var/home/core/zuul-output/0000755000175000017500000000000015114257226014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114263617015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003607651415114263610017710 0ustar rootrootDec 04 10:15:50 crc systemd[1]: Starting Kubernetes Kubelet... Dec 04 10:15:50 crc restorecon[4708]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:50 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 10:15:51 crc restorecon[4708]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 10:15:51 crc restorecon[4708]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 04 10:15:51 crc kubenswrapper[4943]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 10:15:51 crc kubenswrapper[4943]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 04 10:15:51 crc kubenswrapper[4943]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 10:15:51 crc kubenswrapper[4943]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 10:15:51 crc kubenswrapper[4943]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 04 10:15:51 crc kubenswrapper[4943]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.813908 4943 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820454 4943 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820482 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820488 4943 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820494 4943 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820499 4943 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820505 4943 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820511 4943 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820517 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820523 4943 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820529 4943 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820534 4943 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820540 4943 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820545 4943 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820550 4943 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820555 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820560 4943 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820566 4943 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820571 4943 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820576 4943 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820582 4943 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820587 4943 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820593 4943 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820598 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820603 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820608 4943 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820613 4943 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820619 4943 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820636 4943 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820642 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820647 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820653 4943 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820660 4943 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820668 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820675 4943 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820681 4943 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820689 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820695 4943 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820701 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820706 4943 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820711 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820717 4943 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820723 4943 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820728 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820734 4943 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820739 4943 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820744 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820749 4943 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820755 4943 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820761 4943 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820766 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820772 4943 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820777 4943 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820782 4943 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820788 4943 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820794 4943 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820801 4943 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820808 4943 feature_gate.go:330] unrecognized feature gate: Example Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820815 4943 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820821 4943 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820827 4943 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820833 4943 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820839 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820848 4943 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820855 4943 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820861 4943 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820867 4943 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820873 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820879 4943 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820884 4943 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820890 4943 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.820896 4943 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821330 4943 flags.go:64] FLAG: --address="0.0.0.0" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821353 4943 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821370 4943 flags.go:64] FLAG: --anonymous-auth="true" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821381 4943 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821390 4943 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821398 4943 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821409 4943 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821418 4943 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821427 4943 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821434 4943 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821443 4943 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821452 4943 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821459 4943 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821468 4943 flags.go:64] FLAG: --cgroup-root="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821475 4943 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821483 4943 flags.go:64] FLAG: --client-ca-file="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821490 4943 flags.go:64] FLAG: --cloud-config="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821497 4943 flags.go:64] FLAG: --cloud-provider="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821504 4943 flags.go:64] FLAG: --cluster-dns="[]" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821514 4943 flags.go:64] FLAG: --cluster-domain="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821521 4943 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821529 4943 flags.go:64] FLAG: --config-dir="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821537 4943 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821545 4943 flags.go:64] FLAG: --container-log-max-files="5" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821555 4943 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821563 4943 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821571 4943 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821579 4943 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821588 4943 flags.go:64] FLAG: --contention-profiling="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821595 4943 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821603 4943 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821611 4943 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821619 4943 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821629 4943 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821638 4943 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821646 4943 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821653 4943 flags.go:64] FLAG: --enable-load-reader="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821661 4943 flags.go:64] FLAG: --enable-server="true" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821670 4943 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821682 4943 flags.go:64] FLAG: --event-burst="100" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821690 4943 flags.go:64] FLAG: --event-qps="50" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821698 4943 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821706 4943 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821714 4943 flags.go:64] FLAG: --eviction-hard="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821725 4943 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821732 4943 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821740 4943 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821774 4943 flags.go:64] FLAG: --eviction-soft="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821784 4943 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821793 4943 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821801 4943 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821810 4943 flags.go:64] FLAG: --experimental-mounter-path="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821818 4943 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821826 4943 flags.go:64] FLAG: --fail-swap-on="true" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821834 4943 flags.go:64] FLAG: --feature-gates="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821845 4943 flags.go:64] FLAG: --file-check-frequency="20s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821853 4943 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821862 4943 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821871 4943 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821879 4943 flags.go:64] FLAG: --healthz-port="10248" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821887 4943 flags.go:64] FLAG: --help="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821895 4943 flags.go:64] FLAG: --hostname-override="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821903 4943 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821912 4943 flags.go:64] FLAG: --http-check-frequency="20s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821920 4943 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821929 4943 flags.go:64] FLAG: --image-credential-provider-config="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821936 4943 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821944 4943 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821952 4943 flags.go:64] FLAG: --image-service-endpoint="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821959 4943 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821968 4943 flags.go:64] FLAG: --kube-api-burst="100" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821975 4943 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821983 4943 flags.go:64] FLAG: --kube-api-qps="50" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821991 4943 flags.go:64] FLAG: --kube-reserved="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.821998 4943 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822006 4943 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822014 4943 flags.go:64] FLAG: --kubelet-cgroups="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822021 4943 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822028 4943 flags.go:64] FLAG: --lock-file="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822036 4943 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822044 4943 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822052 4943 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822088 4943 flags.go:64] FLAG: --log-json-split-stream="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822098 4943 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822106 4943 flags.go:64] FLAG: --log-text-split-stream="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822113 4943 flags.go:64] FLAG: --logging-format="text" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822121 4943 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822130 4943 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822137 4943 flags.go:64] FLAG: --manifest-url="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822145 4943 flags.go:64] FLAG: --manifest-url-header="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822156 4943 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822164 4943 flags.go:64] FLAG: --max-open-files="1000000" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822174 4943 flags.go:64] FLAG: --max-pods="110" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822182 4943 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822189 4943 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822196 4943 flags.go:64] FLAG: --memory-manager-policy="None" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822233 4943 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822243 4943 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822250 4943 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822258 4943 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822304 4943 flags.go:64] FLAG: --node-status-max-images="50" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822312 4943 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822321 4943 flags.go:64] FLAG: --oom-score-adj="-999" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822329 4943 flags.go:64] FLAG: --pod-cidr="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822337 4943 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822348 4943 flags.go:64] FLAG: --pod-manifest-path="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822356 4943 flags.go:64] FLAG: --pod-max-pids="-1" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822364 4943 flags.go:64] FLAG: --pods-per-core="0" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822372 4943 flags.go:64] FLAG: --port="10250" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822381 4943 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822388 4943 flags.go:64] FLAG: --provider-id="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822396 4943 flags.go:64] FLAG: --qos-reserved="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822405 4943 flags.go:64] FLAG: --read-only-port="10255" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822413 4943 flags.go:64] FLAG: --register-node="true" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822421 4943 flags.go:64] FLAG: --register-schedulable="true" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822429 4943 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822443 4943 flags.go:64] FLAG: --registry-burst="10" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822450 4943 flags.go:64] FLAG: --registry-qps="5" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822458 4943 flags.go:64] FLAG: --reserved-cpus="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822468 4943 flags.go:64] FLAG: --reserved-memory="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822478 4943 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822486 4943 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822494 4943 flags.go:64] FLAG: --rotate-certificates="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822502 4943 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822510 4943 flags.go:64] FLAG: --runonce="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822517 4943 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822527 4943 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822536 4943 flags.go:64] FLAG: --seccomp-default="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822543 4943 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822551 4943 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822559 4943 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822567 4943 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822575 4943 flags.go:64] FLAG: --storage-driver-password="root" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822583 4943 flags.go:64] FLAG: --storage-driver-secure="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822590 4943 flags.go:64] FLAG: --storage-driver-table="stats" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822598 4943 flags.go:64] FLAG: --storage-driver-user="root" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822606 4943 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822614 4943 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822622 4943 flags.go:64] FLAG: --system-cgroups="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822630 4943 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822644 4943 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822651 4943 flags.go:64] FLAG: --tls-cert-file="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822659 4943 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822669 4943 flags.go:64] FLAG: --tls-min-version="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822677 4943 flags.go:64] FLAG: --tls-private-key-file="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822685 4943 flags.go:64] FLAG: --topology-manager-policy="none" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822692 4943 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822700 4943 flags.go:64] FLAG: --topology-manager-scope="container" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822708 4943 flags.go:64] FLAG: --v="2" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822718 4943 flags.go:64] FLAG: --version="false" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822727 4943 flags.go:64] FLAG: --vmodule="" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822737 4943 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.822745 4943 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.822921 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.822933 4943 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.822942 4943 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.822950 4943 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.822957 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.822965 4943 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.822971 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.822978 4943 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.822984 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.822991 4943 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.822999 4943 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823006 4943 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823013 4943 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823020 4943 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823026 4943 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823033 4943 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823045 4943 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823051 4943 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823057 4943 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823064 4943 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823070 4943 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823077 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823083 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823090 4943 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823098 4943 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823104 4943 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823114 4943 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823123 4943 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823130 4943 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823137 4943 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823142 4943 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823148 4943 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823155 4943 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823160 4943 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823167 4943 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823173 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823179 4943 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823185 4943 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823192 4943 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823226 4943 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823233 4943 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823238 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823245 4943 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823250 4943 feature_gate.go:330] unrecognized feature gate: Example Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823256 4943 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823261 4943 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823266 4943 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823271 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823280 4943 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823287 4943 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823293 4943 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823302 4943 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823309 4943 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823316 4943 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823322 4943 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823329 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823343 4943 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823351 4943 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823360 4943 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823367 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823374 4943 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823381 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823387 4943 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823394 4943 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823401 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823407 4943 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823414 4943 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823420 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823427 4943 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823434 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.823441 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.823742 4943 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.830962 4943 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.831010 4943 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831092 4943 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831108 4943 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831114 4943 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831120 4943 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831127 4943 feature_gate.go:330] unrecognized feature gate: Example Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831132 4943 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831138 4943 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831144 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831152 4943 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831163 4943 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831170 4943 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831177 4943 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831184 4943 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831190 4943 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831221 4943 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831228 4943 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831236 4943 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831245 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831252 4943 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831257 4943 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831263 4943 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831268 4943 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831273 4943 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831279 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831284 4943 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831289 4943 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831295 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831300 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831305 4943 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831310 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831317 4943 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831325 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831330 4943 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831336 4943 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831342 4943 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831348 4943 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831353 4943 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831358 4943 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831364 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831369 4943 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831374 4943 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831380 4943 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831385 4943 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831390 4943 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831396 4943 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831401 4943 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831406 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831412 4943 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831417 4943 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831422 4943 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831427 4943 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831432 4943 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831438 4943 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831443 4943 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831448 4943 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831454 4943 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831459 4943 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831466 4943 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831472 4943 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831477 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831483 4943 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831488 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831493 4943 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831498 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831503 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831509 4943 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831514 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831519 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831524 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831529 4943 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831535 4943 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.831544 4943 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831697 4943 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831706 4943 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831712 4943 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831718 4943 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831724 4943 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831730 4943 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831737 4943 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831742 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831749 4943 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831757 4943 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831764 4943 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831770 4943 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831777 4943 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831784 4943 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831790 4943 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831796 4943 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831802 4943 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831808 4943 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831814 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831823 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831828 4943 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831834 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831840 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831845 4943 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831851 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831856 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831861 4943 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831868 4943 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831874 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831880 4943 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831886 4943 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831892 4943 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831897 4943 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831902 4943 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831908 4943 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831914 4943 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831919 4943 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831924 4943 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831930 4943 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831935 4943 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831940 4943 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831945 4943 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831951 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831956 4943 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831962 4943 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831967 4943 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831973 4943 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831980 4943 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831986 4943 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831993 4943 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.831999 4943 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832005 4943 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832011 4943 feature_gate.go:330] unrecognized feature gate: Example Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832016 4943 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832021 4943 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832027 4943 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832033 4943 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832038 4943 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832045 4943 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832051 4943 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832057 4943 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832064 4943 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832069 4943 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832074 4943 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832080 4943 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832085 4943 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832090 4943 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832095 4943 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832102 4943 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832108 4943 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.832114 4943 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.832123 4943 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.832387 4943 server.go:940] "Client rotation is on, will bootstrap in background" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.835647 4943 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.835793 4943 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.837238 4943 server.go:997] "Starting client certificate rotation" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.837288 4943 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.846640 4943 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-05 01:29:41.418502377 +0000 UTC Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.846759 4943 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 15h13m49.571745677s for next certificate rotation Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.868312 4943 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.869942 4943 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.885310 4943 log.go:25] "Validated CRI v1 runtime API" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.912901 4943 log.go:25] "Validated CRI v1 image API" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.915441 4943 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.918643 4943 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-04-10-11-32-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.918755 4943 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.948279 4943 manager.go:217] Machine: {Timestamp:2025-12-04 10:15:51.946356019 +0000 UTC m=+0.535331927 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:32a42f7c-963d-4042-9da3-416ccee17477 BootID:caef8e39-80d0-412e-864d-57fe46732950 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:c6:f2:89 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:c6:f2:89 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:11:36:cc Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:38:ad:be Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:13:0e:5c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:aa:d2:47 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:16:f4:43:f6:d2:1e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:56:5d:13:46:78:57 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.948726 4943 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.949040 4943 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.949609 4943 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.949937 4943 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.950000 4943 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.950350 4943 topology_manager.go:138] "Creating topology manager with none policy" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.950368 4943 container_manager_linux.go:303] "Creating device plugin manager" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.950778 4943 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.950867 4943 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.951326 4943 state_mem.go:36] "Initialized new in-memory state store" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.951470 4943 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.952706 4943 kubelet.go:418] "Attempting to sync node with API server" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.952733 4943 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.952794 4943 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.952834 4943 kubelet.go:324] "Adding apiserver pod source" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.952877 4943 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.963730 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.963842 4943 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 04 10:15:51 crc kubenswrapper[4943]: E1204 10:15:51.963869 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 04 10:15:51 crc kubenswrapper[4943]: W1204 10:15:51.964185 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.971287 4943 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.972366 4943 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973120 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973156 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973170 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973181 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973215 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973228 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973240 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973257 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973271 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973284 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973299 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973307 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.973504 4943 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.974049 4943 server.go:1280] "Started kubelet" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.974415 4943 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 04 10:15:51 crc systemd[1]: Started Kubernetes Kubelet. Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.988569 4943 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.989428 4943 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 04 10:15:51 crc kubenswrapper[4943]: E1204 10:15:51.989788 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.990792 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.990882 4943 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.991055 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 01:42:28.059394931 +0000 UTC Dec 04 10:15:51 crc kubenswrapper[4943]: E1204 10:15:51.990746 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187dfbad969eba55 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 10:15:51.974001237 +0000 UTC m=+0.562977105,LastTimestamp:2025-12-04 10:15:51.974001237 +0000 UTC m=+0.562977105,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.991334 4943 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.991371 4943 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 04 10:15:51 crc kubenswrapper[4943]: E1204 10:15:51.991483 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 10:15:51 crc kubenswrapper[4943]: E1204 10:15:51.992116 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="200ms" Dec 04 10:15:51 crc kubenswrapper[4943]: I1204 10:15:51.992290 4943 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.403187 4943 factory.go:55] Registering systemd factory Dec 04 10:15:52 crc kubenswrapper[4943]: W1204 10:15:52.407943 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:52 crc kubenswrapper[4943]: E1204 10:15:52.408806 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 10:15:52 crc kubenswrapper[4943]: E1204 10:15:52.409326 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.409337 4943 factory.go:221] Registration of the systemd container factory successfully Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.410050 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.411183 4943 factory.go:153] Registering CRI-O factory Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.411228 4943 factory.go:221] Registration of the crio container factory successfully Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.411311 4943 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.411351 4943 factory.go:103] Registering Raw factory Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.411395 4943 manager.go:1196] Started watching for new ooms in manager Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.412336 4943 server.go:460] "Adding debug handlers to kubelet server" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.412339 4943 manager.go:319] Starting recovery of all containers Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.418004 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.418156 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.418261 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.418343 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.418437 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.418522 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.418617 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.418718 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.418815 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.418907 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.418989 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.419102 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.419193 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.419321 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: E1204 10:15:52.413576 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="400ms" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.419430 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.419623 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.419699 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.419772 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.421228 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.423165 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.423302 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.423391 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.423473 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.423564 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.423688 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.423778 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.423869 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.423955 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.424074 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.424169 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.424268 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.424377 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.424467 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.424547 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.424634 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.424752 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.424832 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.424895 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.424970 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.425049 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.425121 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.425180 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.425376 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.425465 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.425564 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.425658 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.425754 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.425842 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.425938 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.426029 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.426126 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.426230 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.426332 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.426438 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.426516 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.426612 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.426710 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.426800 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.426883 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.426979 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.427069 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.427166 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.427281 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.427378 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.427482 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.427567 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.427765 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.427868 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.427964 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.428056 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.428145 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.428274 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.428365 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.428448 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.428529 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.428618 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.428711 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.428806 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.428904 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.429005 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.429179 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.429326 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.429419 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.429528 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.429608 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.429684 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.429771 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.429862 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.429941 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.430215 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.430300 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.430394 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.430473 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.430566 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.430671 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.430756 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.430867 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.430955 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.431227 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.431315 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.431400 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.431483 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.431608 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.431707 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.431799 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.431897 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.431975 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.432051 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.432151 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.432264 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.432358 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.432447 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.432551 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.432654 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.432748 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.432832 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.432911 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.432993 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.433085 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.433172 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.433303 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.433389 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.433487 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.433588 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.433675 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.433763 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.433846 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.432657 4943 manager.go:324] Recovery completed Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.433930 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.434128 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.434231 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.434317 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.434404 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.434494 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.434576 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.434668 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.434896 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.434991 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.435079 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.435148 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.435237 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.435332 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.435408 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.435487 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.435595 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.435682 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.435771 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.435863 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.435951 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.436048 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.436157 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.436270 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.436365 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.436447 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.436521 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.436584 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.436641 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.436705 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.436770 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.436883 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.436963 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.437291 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.437495 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.437582 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.437661 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.437742 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.437816 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.437900 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.437982 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.438059 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.438325 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.438412 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.438506 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.438602 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.438684 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.438768 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.438856 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.438932 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.439028 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.439116 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.439215 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.439306 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.439397 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.439484 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.439592 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.439677 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.439757 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.439830 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.439916 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.440105 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.440217 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.440306 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.440401 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.440503 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.441291 4943 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.441418 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.441507 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.441592 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.441685 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.441778 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.441890 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.446824 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.447167 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.447284 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.447421 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.447516 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.447637 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.447715 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.447806 4943 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.447890 4943 reconstruct.go:97] "Volume reconstruction finished" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.447963 4943 reconciler.go:26] "Reconciler: start to sync state" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.451854 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.454605 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.454646 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.454657 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.458513 4943 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.458540 4943 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.458567 4943 state_mem.go:36] "Initialized new in-memory state store" Dec 04 10:15:52 crc kubenswrapper[4943]: E1204 10:15:52.510382 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.560638 4943 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.561622 4943 policy_none.go:49] "None policy: Start" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.562869 4943 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.562913 4943 state_mem.go:35] "Initializing new in-memory state store" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.563215 4943 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.563331 4943 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.563407 4943 kubelet.go:2335] "Starting kubelet main sync loop" Dec 04 10:15:52 crc kubenswrapper[4943]: E1204 10:15:52.563564 4943 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 04 10:15:52 crc kubenswrapper[4943]: W1204 10:15:52.564696 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:52 crc kubenswrapper[4943]: E1204 10:15:52.564795 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 04 10:15:52 crc kubenswrapper[4943]: E1204 10:15:52.610726 4943 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.632591 4943 manager.go:334] "Starting Device Plugin manager" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.632651 4943 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.632665 4943 server.go:79] "Starting device plugin registration server" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.633094 4943 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.633127 4943 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.633309 4943 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.633541 4943 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.633555 4943 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 04 10:15:52 crc kubenswrapper[4943]: E1204 10:15:52.641373 4943 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.664085 4943 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.664242 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.665714 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.665809 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.665884 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.666151 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.666488 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.666550 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.667256 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.667303 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.667314 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.667517 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.667595 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.667632 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.667641 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.667783 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.667880 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.668799 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.668839 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.668852 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.669688 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.669759 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.669777 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.670073 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.670281 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.670372 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.671403 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.671648 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.671713 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.671801 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.671817 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.671913 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.672048 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.672307 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.672364 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.673894 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.673983 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.674050 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.674134 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.674173 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.674189 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.674504 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.674545 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.676022 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.676048 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.676063 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.734226 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.735903 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.735943 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.735953 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.735987 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 10:15:52 crc kubenswrapper[4943]: E1204 10:15:52.736784 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.762484 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.762538 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.762567 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.762593 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.762726 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.762759 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.762860 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.762945 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.762972 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.762996 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.763016 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.763043 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.763065 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.763105 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.763139 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: E1204 10:15:52.820541 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="800ms" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.864734 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.864788 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.864807 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.864823 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.864838 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.864940 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.864856 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.864988 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865031 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865020 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865011 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865032 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865054 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865157 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865035 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865252 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865191 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865303 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865311 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865372 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865400 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865405 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865430 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865285 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865480 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865510 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865541 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865592 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865606 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.865673 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.936928 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.938437 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.938470 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.938480 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.938503 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 10:15:52 crc kubenswrapper[4943]: E1204 10:15:52.938971 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 04 10:15:52 crc kubenswrapper[4943]: W1204 10:15:52.967437 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:52 crc kubenswrapper[4943]: E1204 10:15:52.967558 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.992041 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 22:36:46.291279634 +0000 UTC Dec 04 10:15:52 crc kubenswrapper[4943]: I1204 10:15:52.995765 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.002304 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 10:15:53 crc kubenswrapper[4943]: W1204 10:15:53.023526 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-8379910328fb393ca25943f08683e4bbf00f7cef2be87cc6db90b0cdbe7735e8 WatchSource:0}: Error finding container 8379910328fb393ca25943f08683e4bbf00f7cef2be87cc6db90b0cdbe7735e8: Status 404 returned error can't find the container with id 8379910328fb393ca25943f08683e4bbf00f7cef2be87cc6db90b0cdbe7735e8 Dec 04 10:15:53 crc kubenswrapper[4943]: W1204 10:15:53.026755 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-d92656ba19bfbd2fd2c7b5885e9b60b532ac32e8f48d625511c51f5073ee018d WatchSource:0}: Error finding container d92656ba19bfbd2fd2c7b5885e9b60b532ac32e8f48d625511c51f5073ee018d: Status 404 returned error can't find the container with id d92656ba19bfbd2fd2c7b5885e9b60b532ac32e8f48d625511c51f5073ee018d Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.040914 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:53 crc kubenswrapper[4943]: W1204 10:15:53.054647 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-4c359e518dfac9dc6aa116b56aa99209c5cc2a9b63a8d3c7288989d2d405d59f WatchSource:0}: Error finding container 4c359e518dfac9dc6aa116b56aa99209c5cc2a9b63a8d3c7288989d2d405d59f: Status 404 returned error can't find the container with id 4c359e518dfac9dc6aa116b56aa99209c5cc2a9b63a8d3c7288989d2d405d59f Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.066402 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.072417 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 10:15:53 crc kubenswrapper[4943]: W1204 10:15:53.091110 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-f23a2ab7f5f0412cba95f9449f7a6befb9efb659493168ff7c488391858bd37a WatchSource:0}: Error finding container f23a2ab7f5f0412cba95f9449f7a6befb9efb659493168ff7c488391858bd37a: Status 404 returned error can't find the container with id f23a2ab7f5f0412cba95f9449f7a6befb9efb659493168ff7c488391858bd37a Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.339895 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.341904 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.341948 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.341959 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.341980 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 10:15:53 crc kubenswrapper[4943]: E1204 10:15:53.342373 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.412413 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:53 crc kubenswrapper[4943]: W1204 10:15:53.496248 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:53 crc kubenswrapper[4943]: E1204 10:15:53.496351 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 04 10:15:53 crc kubenswrapper[4943]: W1204 10:15:53.516214 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:53 crc kubenswrapper[4943]: E1204 10:15:53.516271 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.567633 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"37dccb547a1bfa50ade37103cb52fa5698dba171e54d2236aa513ffa12776ad8"} Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.569012 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4c359e518dfac9dc6aa116b56aa99209c5cc2a9b63a8d3c7288989d2d405d59f"} Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.570026 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8379910328fb393ca25943f08683e4bbf00f7cef2be87cc6db90b0cdbe7735e8"} Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.571127 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"d92656ba19bfbd2fd2c7b5885e9b60b532ac32e8f48d625511c51f5073ee018d"} Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.572114 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f23a2ab7f5f0412cba95f9449f7a6befb9efb659493168ff7c488391858bd37a"} Dec 04 10:15:53 crc kubenswrapper[4943]: E1204 10:15:53.621964 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="1.6s" Dec 04 10:15:53 crc kubenswrapper[4943]: W1204 10:15:53.976464 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:53 crc kubenswrapper[4943]: E1204 10:15:53.976975 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 04 10:15:53 crc kubenswrapper[4943]: I1204 10:15:53.992349 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 03:35:53.898772811 +0000 UTC Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.142763 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.144651 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.144716 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.144739 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.144777 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 10:15:54 crc kubenswrapper[4943]: E1204 10:15:54.145688 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 04 10:15:54 crc kubenswrapper[4943]: E1204 10:15:54.199348 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187dfbad969eba55 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 10:15:51.974001237 +0000 UTC m=+0.562977105,LastTimestamp:2025-12-04 10:15:51.974001237 +0000 UTC m=+0.562977105,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.412690 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.581248 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465"} Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.581356 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2"} Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.581383 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3"} Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.581394 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf"} Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.581482 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.582703 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.582730 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.582746 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.585218 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0" exitCode=0 Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.585305 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0"} Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.585447 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.586332 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.586356 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.586367 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.588570 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.589512 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.589543 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.589557 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.589871 4943 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="141e6d35a5a43cc2dcd8b9c2c70d3745a9f53d92178b4ef6718dc776a6ee1a3b" exitCode=0 Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.589980 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"141e6d35a5a43cc2dcd8b9c2c70d3745a9f53d92178b4ef6718dc776a6ee1a3b"} Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.590283 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.591510 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.591538 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.591550 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.592178 4943 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="671f590dfaa03935b9ff581ce09c8c0d9d6e7c7fdd5ca649ec973fa2d94e6f49" exitCode=0 Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.592313 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"671f590dfaa03935b9ff581ce09c8c0d9d6e7c7fdd5ca649ec973fa2d94e6f49"} Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.592401 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.593334 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.593379 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.593419 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.594909 4943 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b" exitCode=0 Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.594969 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b"} Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.595053 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.606534 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.606710 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:54 crc kubenswrapper[4943]: I1204 10:15:54.606738 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.032830 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 00:39:12.649392329 +0000 UTC Dec 04 10:15:55 crc kubenswrapper[4943]: E1204 10:15:55.223427 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="3.2s" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.418472 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.609136 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc"} Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.609259 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465"} Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.609271 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9"} Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.611904 4943 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="18b4a76c7fed98c4f74e31ae34bd4948bea954e2a49ad118b35cbb5b26bc2faf" exitCode=0 Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.612106 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.612073 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"18b4a76c7fed98c4f74e31ae34bd4948bea954e2a49ad118b35cbb5b26bc2faf"} Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.613466 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.613502 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.613515 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.615618 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"2631d397309030146ffbdbe86cfac6e02dd630c8347c6418d19b46300237e50a"} Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.615721 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.617082 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.617104 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.617115 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.622692 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.623306 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.623679 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a61720056d847f2e14a4c17ca79e24d8e330530b0ed68c256aa2ddcb35552c25"} Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.623722 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0e348e3b6d23732341818a2d444fc18b715cc7729d49508ba0bec8a349118aba"} Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.623737 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"07144e00536ba9c9e2f8f9be3199ca05f6b6a864d5ee1fe01b49c4e30a244ae1"} Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.624223 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.624248 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.624259 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.625186 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.625226 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.625236 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:55 crc kubenswrapper[4943]: W1204 10:15:55.711691 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:55 crc kubenswrapper[4943]: E1204 10:15:55.711867 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.746624 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.748855 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.748911 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.749250 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:55 crc kubenswrapper[4943]: I1204 10:15:55.749314 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 10:15:55 crc kubenswrapper[4943]: E1204 10:15:55.750051 4943 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.150:6443: connect: connection refused" node="crc" Dec 04 10:15:55 crc kubenswrapper[4943]: W1204 10:15:55.876819 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.150:6443: connect: connection refused Dec 04 10:15:55 crc kubenswrapper[4943]: E1204 10:15:55.876950 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.150:6443: connect: connection refused" logger="UnhandledError" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.033484 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 15:25:48.616413665 +0000 UTC Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.033553 4943 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 5h9m52.582864227s for next certificate rotation Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.636767 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00"} Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.636823 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf"} Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.636886 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.637632 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.637673 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.637684 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.638687 4943 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9e881c9fc1b5a53b0c3e9e298071be7bdd6f674a3140cc4a6b1d0073b6ff2494" exitCode=0 Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.638764 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.638770 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9e881c9fc1b5a53b0c3e9e298071be7bdd6f674a3140cc4a6b1d0073b6ff2494"} Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.638905 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.638928 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.639264 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.639329 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.639349 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.639357 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.640021 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.640044 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.640059 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.640959 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.640976 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:56 crc kubenswrapper[4943]: I1204 10:15:56.640984 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.237808 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.652053 4943 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.652120 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.652309 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8f18df637c05e1f517e5f5e41d9977e688f153198a7e144494c7ea3d31ed30f5"} Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.652372 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.652435 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.652394 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"48144828027b06c9f247dce0123e1b45080288632288017b26ce1906896822ce"} Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.652514 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fe322eaf4885153ea4ea0e0f059997517b5c4d83d102d4ccd426deedbded72e8"} Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.652543 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3db38524a3da9b4b5a7bf35d9329a0aa6a74d3bd3d7424feebb1cdabc8e043c2"} Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.652557 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"716ef8a0bf82dcbc707232cf9e7a72b09730d7cbd421950b792eb5e04face665"} Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.653563 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.653598 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.653611 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.653576 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.653689 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.653704 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.654402 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.654537 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:57 crc kubenswrapper[4943]: I1204 10:15:57.654555 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.138044 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.138272 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.139513 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.139553 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.139575 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.655450 4943 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.655513 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.655571 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.656818 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.656891 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.656913 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.657014 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.657047 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.657057 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.951054 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.952624 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.952800 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.952976 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:15:58 crc kubenswrapper[4943]: I1204 10:15:58.953094 4943 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 10:15:59 crc kubenswrapper[4943]: I1204 10:15:59.037660 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:15:59 crc kubenswrapper[4943]: I1204 10:15:59.660838 4943 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 10:15:59 crc kubenswrapper[4943]: I1204 10:15:59.662374 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:15:59 crc kubenswrapper[4943]: I1204 10:15:59.664010 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:15:59 crc kubenswrapper[4943]: I1204 10:15:59.664055 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:15:59 crc kubenswrapper[4943]: I1204 10:15:59.664065 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:00 crc kubenswrapper[4943]: I1204 10:16:00.510587 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:16:00 crc kubenswrapper[4943]: I1204 10:16:00.510888 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:16:00 crc kubenswrapper[4943]: I1204 10:16:00.512515 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:00 crc kubenswrapper[4943]: I1204 10:16:00.512546 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:00 crc kubenswrapper[4943]: I1204 10:16:00.512555 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:00 crc kubenswrapper[4943]: I1204 10:16:00.515939 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 04 10:16:00 crc kubenswrapper[4943]: I1204 10:16:00.516126 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:16:00 crc kubenswrapper[4943]: I1204 10:16:00.517431 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:00 crc kubenswrapper[4943]: I1204 10:16:00.517449 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:00 crc kubenswrapper[4943]: I1204 10:16:00.517457 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.516055 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.516335 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.517598 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.517642 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.517651 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.521133 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.615551 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.615824 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.617051 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.617108 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.617125 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.665626 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.665705 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.667232 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.667266 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:01 crc kubenswrapper[4943]: I1204 10:16:01.667278 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:02 crc kubenswrapper[4943]: E1204 10:16:02.655382 4943 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 10:16:02 crc kubenswrapper[4943]: I1204 10:16:02.667902 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:16:02 crc kubenswrapper[4943]: I1204 10:16:02.669063 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:02 crc kubenswrapper[4943]: I1204 10:16:02.669139 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:02 crc kubenswrapper[4943]: I1204 10:16:02.669160 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:02 crc kubenswrapper[4943]: I1204 10:16:02.850601 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:16:02 crc kubenswrapper[4943]: I1204 10:16:02.850847 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:16:02 crc kubenswrapper[4943]: I1204 10:16:02.852541 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:02 crc kubenswrapper[4943]: I1204 10:16:02.852604 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:02 crc kubenswrapper[4943]: I1204 10:16:02.852618 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:03 crc kubenswrapper[4943]: I1204 10:16:03.511823 4943 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 10:16:03 crc kubenswrapper[4943]: I1204 10:16:03.511996 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 10:16:03 crc kubenswrapper[4943]: I1204 10:16:03.590660 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:16:03 crc kubenswrapper[4943]: I1204 10:16:03.670684 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:16:03 crc kubenswrapper[4943]: I1204 10:16:03.671619 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:03 crc kubenswrapper[4943]: I1204 10:16:03.671656 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:03 crc kubenswrapper[4943]: I1204 10:16:03.671668 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:06 crc kubenswrapper[4943]: W1204 10:16:06.194638 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 10:16:06 crc kubenswrapper[4943]: I1204 10:16:06.194770 4943 trace.go:236] Trace[417834907]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 10:15:56.193) (total time: 10000ms): Dec 04 10:16:06 crc kubenswrapper[4943]: Trace[417834907]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (10:16:06.194) Dec 04 10:16:06 crc kubenswrapper[4943]: Trace[417834907]: [10.00095177s] [10.00095177s] END Dec 04 10:16:06 crc kubenswrapper[4943]: E1204 10:16:06.194831 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 10:16:06 crc kubenswrapper[4943]: I1204 10:16:06.413281 4943 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 04 10:16:06 crc kubenswrapper[4943]: W1204 10:16:06.729918 4943 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 04 10:16:06 crc kubenswrapper[4943]: I1204 10:16:06.730041 4943 trace.go:236] Trace[731194582]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 10:15:56.728) (total time: 10001ms): Dec 04 10:16:06 crc kubenswrapper[4943]: Trace[731194582]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (10:16:06.729) Dec 04 10:16:06 crc kubenswrapper[4943]: Trace[731194582]: [10.001421329s] [10.001421329s] END Dec 04 10:16:06 crc kubenswrapper[4943]: E1204 10:16:06.730068 4943 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 10:16:07 crc kubenswrapper[4943]: I1204 10:16:07.238766 4943 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 10:16:07 crc kubenswrapper[4943]: I1204 10:16:07.238883 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 10:16:08 crc kubenswrapper[4943]: I1204 10:16:08.276246 4943 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 04 10:16:08 crc kubenswrapper[4943]: I1204 10:16:08.276420 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 04 10:16:10 crc kubenswrapper[4943]: I1204 10:16:10.541899 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 04 10:16:10 crc kubenswrapper[4943]: I1204 10:16:10.542172 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:16:10 crc kubenswrapper[4943]: I1204 10:16:10.543415 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:10 crc kubenswrapper[4943]: I1204 10:16:10.543476 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:10 crc kubenswrapper[4943]: I1204 10:16:10.543492 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:10 crc kubenswrapper[4943]: I1204 10:16:10.556934 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 04 10:16:10 crc kubenswrapper[4943]: I1204 10:16:10.689394 4943 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 10:16:10 crc kubenswrapper[4943]: I1204 10:16:10.691018 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:10 crc kubenswrapper[4943]: I1204 10:16:10.691101 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:10 crc kubenswrapper[4943]: I1204 10:16:10.691117 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:11 crc kubenswrapper[4943]: I1204 10:16:11.160830 4943 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.036781 4943 apiserver.go:52] "Watching apiserver" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.043578 4943 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.043949 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.044502 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.044611 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:12 crc kubenswrapper[4943]: E1204 10:16:12.044788 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.044850 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.044942 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.044948 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.044972 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:12 crc kubenswrapper[4943]: E1204 10:16:12.045026 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:12 crc kubenswrapper[4943]: E1204 10:16:12.045080 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.046900 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.047561 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.047779 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.048228 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.048298 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.048322 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.048244 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.048918 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.050355 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.072343 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.086810 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.093668 4943 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.097527 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.109421 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.119353 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.129548 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.141999 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.244546 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.248835 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.258310 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.260549 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.268936 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.280528 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.293727 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.305884 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.317734 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.329121 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.339006 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.351156 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.361493 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.372251 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.381853 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.391034 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.588731 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.612625 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.658090 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.675219 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.688448 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.699374 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: E1204 10:16:12.703612 4943 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.713831 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:12 crc kubenswrapper[4943]: I1204 10:16:12.754641 4943 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.260351 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.261677 4943 trace.go:236] Trace[2001641919]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 10:16:00.165) (total time: 13095ms): Dec 04 10:16:13 crc kubenswrapper[4943]: Trace[2001641919]: ---"Objects listed" error: 13095ms (10:16:13.261) Dec 04 10:16:13 crc kubenswrapper[4943]: Trace[2001641919]: [13.095662631s] [13.095662631s] END Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.261711 4943 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.268791 4943 trace.go:236] Trace[159135786]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 10:15:59.155) (total time: 14113ms): Dec 04 10:16:13 crc kubenswrapper[4943]: Trace[159135786]: ---"Objects listed" error: 14113ms (10:16:13.268) Dec 04 10:16:13 crc kubenswrapper[4943]: Trace[159135786]: [14.113648423s] [14.113648423s] END Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.268820 4943 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.302604 4943 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.403573 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.403732 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.403799 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.403852 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.403939 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.403987 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404049 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404069 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404094 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404099 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404184 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404309 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404371 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404430 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404479 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404486 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404526 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404546 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404568 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404585 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404605 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404624 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404642 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404670 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404684 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404701 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404734 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404751 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404762 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404767 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404830 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404854 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404901 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404903 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404934 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404952 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404969 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.404986 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405011 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405050 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405067 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405088 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405122 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405113 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405138 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405236 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405262 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405283 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405292 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405301 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405323 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405607 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405786 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405847 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405857 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405931 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405942 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.405947 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.406017 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.406123 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.406145 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.406223 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.406682 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.406782 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.406853 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.406920 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.406986 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407107 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407255 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407344 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407414 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407487 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407563 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407646 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407719 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407795 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407869 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407971 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408038 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408106 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408182 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408263 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408360 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408437 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408504 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408581 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408655 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408723 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.406687 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.406938 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407049 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407239 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407406 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407466 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407632 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407649 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407739 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.407561 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408374 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408491 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408781 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.408911 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.409031 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.409034 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.411772 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.412400 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.412493 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.412564 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.412629 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.412696 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.412761 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.412829 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.412895 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413045 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413120 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413187 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413292 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413396 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413465 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413536 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413606 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413683 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413948 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.414017 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.414085 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.414152 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.414984 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.428993 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429089 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429135 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429163 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429219 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429239 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429276 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429299 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429322 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429393 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429417 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429438 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429485 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429519 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429576 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429601 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429649 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429681 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429728 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429757 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430024 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430062 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430113 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430149 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430226 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430256 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430317 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430352 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430407 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430438 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430549 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430583 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430628 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430660 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430705 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430733 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430783 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430812 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430859 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430885 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430913 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430965 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430990 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431043 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431097 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431130 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431182 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431248 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431309 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431343 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431401 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431440 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431495 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431549 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431598 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431670 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431727 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431765 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431820 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431850 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431901 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431933 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431993 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432063 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432166 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432229 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432262 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432353 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432404 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432438 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432502 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432563 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432598 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432661 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432722 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432752 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432810 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432845 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432896 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432949 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432982 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433032 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433064 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433115 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433152 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433215 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433248 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433299 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433324 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433377 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433409 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433457 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433489 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433545 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433606 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433636 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433685 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433718 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433764 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433799 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433855 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433887 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433933 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433964 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434104 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434179 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434253 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434315 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434398 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434466 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434505 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.412528 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.412798 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413023 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413088 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413250 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.413657 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.414006 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.414532 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.414861 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.414926 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.414994 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.415019 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429069 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429468 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.429628 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430910 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.430960 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431039 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431109 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431429 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.431656 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432517 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.432119 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.433574 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434004 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434184 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434360 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434504 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434753 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.436703 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.436761 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.437762 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.438032 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.438230 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.438469 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.438893 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.439996 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.439972 4943 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.440628 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.440981 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.441276 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.441520 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.442349 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.443058 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.444187 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.444999 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.447256 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.447399 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.447826 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.450810 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.453982 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.454544 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.456071 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.456376 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.456603 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.456824 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.457336 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.457690 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.458011 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.458060 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.458513 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.458463 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.459070 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.459522 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.459670 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.459747 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.459758 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.459900 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.459944 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.460178 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:13.9601498 +0000 UTC m=+22.549125668 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.460192 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.460490 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.460520 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.460761 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.461081 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.461213 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.461368 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.461513 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.461520 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.461562 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.461722 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.461819 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.461971 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.462155 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.462534 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.462641 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.462657 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.462666 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.463041 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.463072 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.462987 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.463713 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.463849 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.463918 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.464775 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.465461 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.466277 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.466444 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.466659 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.466839 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.467111 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.494731 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.494990 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.495168 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.495393 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.495451 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.495559 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:16:13.995533417 +0000 UTC m=+22.584509285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.496131 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.496308 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.497034 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.497063 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.497482 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.498056 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.498214 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.498282 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.498351 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.498555 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.498784 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.498874 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.499015 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.499036 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.499158 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.499394 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.499533 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.499723 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.500056 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.500188 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.500467 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.503258 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.503734 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.504063 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.504256 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.504454 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.504736 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.505301 4943 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.505499 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.505531 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.505543 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.505568 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.505612 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:14.005594099 +0000 UTC m=+22.594569967 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.434739 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.505826 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.505860 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.505897 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.505924 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.505947 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.505974 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506074 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506090 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506103 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506115 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506126 4943 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506140 4943 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506163 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506179 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506192 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506226 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506238 4943 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506249 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506260 4943 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506272 4943 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506283 4943 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506296 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506307 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506318 4943 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506326 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506335 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506345 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506358 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506365 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506376 4943 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506384 4943 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506393 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506408 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506419 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506429 4943 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506437 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506445 4943 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506454 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506461 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506470 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506478 4943 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506486 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506494 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.506502 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.505560 4943 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.507366 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508378 4943 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508399 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508413 4943 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508430 4943 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508444 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508457 4943 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508472 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508485 4943 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508500 4943 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508513 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508529 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508547 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508562 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508579 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508592 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508606 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508618 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508631 4943 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508688 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508702 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508713 4943 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508729 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508742 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508756 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508769 4943 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508783 4943 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508797 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508809 4943 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508822 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508833 4943 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508850 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508863 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508880 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508893 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508907 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508919 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508936 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508967 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.508991 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509004 4943 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509018 4943 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509031 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509043 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509058 4943 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509072 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509088 4943 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509107 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509125 4943 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509138 4943 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509151 4943 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509166 4943 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509180 4943 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509193 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509228 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509239 4943 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509253 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509264 4943 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509279 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509291 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509305 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509318 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509331 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509344 4943 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509357 4943 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509368 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509380 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509391 4943 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509404 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509415 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509428 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509448 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509463 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509477 4943 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509491 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509507 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509520 4943 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509534 4943 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509548 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509561 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509575 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509588 4943 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509601 4943 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509614 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509628 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509642 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509655 4943 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509667 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509681 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509694 4943 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509706 4943 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509716 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509727 4943 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509740 4943 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509756 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509767 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509779 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509794 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509809 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509821 4943 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509834 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509846 4943 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509859 4943 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509878 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509891 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509902 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509915 4943 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509927 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509945 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509959 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509982 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.509997 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510012 4943 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510027 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510040 4943 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510053 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510066 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510079 4943 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510093 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510104 4943 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510118 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510131 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510143 4943 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510156 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510170 4943 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510182 4943 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510213 4943 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510227 4943 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.510247 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.511376 4943 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.511428 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.511732 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.511752 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.511761 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.511787 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.511813 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:13Z","lastTransitionTime":"2025-12-04T10:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.513451 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.513679 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.513727 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.513761 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.513897 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:14.013871568 +0000 UTC m=+22.602847436 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.514126 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.514325 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.514503 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.514993 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.515318 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.515496 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.515766 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.515909 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.517301 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.517819 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.518296 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.521309 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.522475 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.527445 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.528376 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.530118 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.530602 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.531294 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.531404 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.537430 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.538755 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.541086 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.541886 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.543738 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.545576 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.545589 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.545762 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.545775 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.545778 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.548294 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.564539 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.564696 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.564766 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.564819 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.564915 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.565059 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.568052 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.568313 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.574739 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.591567 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610695 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610738 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610797 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610811 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610821 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610831 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610842 4943 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610853 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610864 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610878 4943 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610889 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610900 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610912 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610924 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610936 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610945 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610953 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610962 4943 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610969 4943 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610977 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610985 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.610993 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.611003 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.611013 4943 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.611021 4943 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.611029 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.611036 4943 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.611044 4943 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.611052 4943 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.611099 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.611283 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.624264 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.625053 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.625176 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.625283 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.625419 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:14.125397831 +0000 UTC m=+22.714373699 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.627560 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.633926 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.633961 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.633972 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.633990 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.634002 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:13Z","lastTransitionTime":"2025-12-04T10:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.657153 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.661071 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.662490 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.662557 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.662569 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.662585 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.662596 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:13Z","lastTransitionTime":"2025-12-04T10:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.677951 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.678000 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.682721 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.682774 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.682786 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.682804 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.682816 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:13Z","lastTransitionTime":"2025-12-04T10:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.690815 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.692243 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.695923 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6a555511fa1fd6cf009166259581c7791a7b0775b16191d3667146ebc002ebbf"} Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.698603 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.698670 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.698690 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.698710 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.698723 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:13Z","lastTransitionTime":"2025-12-04T10:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.703447 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.713726 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: E1204 10:16:13.713846 4943 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.715373 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.718243 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.718308 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.718323 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.718353 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.718368 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:13Z","lastTransitionTime":"2025-12-04T10:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.729960 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.778805 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-wkbf5"] Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.779127 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-lfcbx"] Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.779246 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-kswzd"] Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.779477 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.779767 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.780073 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lfcbx" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.781421 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.781620 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.781835 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.782016 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.783593 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.784257 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.784397 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.785117 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.785648 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.785761 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.785778 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.785899 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.786470 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.803323 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.811870 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501-proxy-tls\") pod \"machine-config-daemon-kswzd\" (UID: \"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\") " pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.811912 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-run-netns\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.811928 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-multus-daemon-config\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.811944 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-etc-kubernetes\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.811968 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w667f\" (UniqueName: \"kubernetes.io/projected/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-kube-api-access-w667f\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.811984 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjxng\" (UniqueName: \"kubernetes.io/projected/a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc-kube-api-access-hjxng\") pod \"node-resolver-lfcbx\" (UID: \"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\") " pod="openshift-dns/node-resolver-lfcbx" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812417 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-hostroot\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812441 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-run-multus-certs\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812457 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc-hosts-file\") pod \"node-resolver-lfcbx\" (UID: \"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\") " pod="openshift-dns/node-resolver-lfcbx" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812473 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbwhk\" (UniqueName: \"kubernetes.io/projected/7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501-kube-api-access-qbwhk\") pod \"machine-config-daemon-kswzd\" (UID: \"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\") " pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812490 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-system-cni-dir\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812506 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-var-lib-cni-multus\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812523 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-var-lib-kubelet\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812570 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501-rootfs\") pod \"machine-config-daemon-kswzd\" (UID: \"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\") " pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812587 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-cni-binary-copy\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812604 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-var-lib-cni-bin\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812623 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-multus-conf-dir\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812660 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-cnibin\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812725 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-os-release\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812773 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-multus-socket-dir-parent\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812807 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501-mcd-auth-proxy-config\") pod \"machine-config-daemon-kswzd\" (UID: \"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\") " pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812828 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-multus-cni-dir\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.812854 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-run-k8s-cni-cncf-io\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.823301 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.823370 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.823408 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.823418 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.823441 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.823453 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:13Z","lastTransitionTime":"2025-12-04T10:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.837416 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.849761 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.858285 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.871000 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: W1204 10:16:13.872032 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-d7dbc4b292bd4ff5b1b50087184766e52a698412f33d771c2db9770191825f89 WatchSource:0}: Error finding container d7dbc4b292bd4ff5b1b50087184766e52a698412f33d771c2db9770191825f89: Status 404 returned error can't find the container with id d7dbc4b292bd4ff5b1b50087184766e52a698412f33d771c2db9770191825f89 Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.876595 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.882018 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.896786 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.907244 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.913566 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501-mcd-auth-proxy-config\") pod \"machine-config-daemon-kswzd\" (UID: \"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\") " pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.913784 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-multus-cni-dir\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.913859 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-run-k8s-cni-cncf-io\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.913927 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501-proxy-tls\") pod \"machine-config-daemon-kswzd\" (UID: \"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\") " pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914014 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-run-netns\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914188 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-multus-daemon-config\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914338 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-etc-kubernetes\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914437 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjxng\" (UniqueName: \"kubernetes.io/projected/a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc-kube-api-access-hjxng\") pod \"node-resolver-lfcbx\" (UID: \"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\") " pod="openshift-dns/node-resolver-lfcbx" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914498 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501-mcd-auth-proxy-config\") pod \"machine-config-daemon-kswzd\" (UID: \"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\") " pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914536 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-run-k8s-cni-cncf-io\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914569 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-multus-cni-dir\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914513 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w667f\" (UniqueName: \"kubernetes.io/projected/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-kube-api-access-w667f\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914261 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-run-netns\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914632 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-etc-kubernetes\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914677 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-system-cni-dir\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914724 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-var-lib-cni-multus\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914746 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-hostroot\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914767 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-run-multus-certs\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914787 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc-hosts-file\") pod \"node-resolver-lfcbx\" (UID: \"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\") " pod="openshift-dns/node-resolver-lfcbx" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914807 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbwhk\" (UniqueName: \"kubernetes.io/projected/7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501-kube-api-access-qbwhk\") pod \"machine-config-daemon-kswzd\" (UID: \"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\") " pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914831 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-var-lib-kubelet\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914847 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-hostroot\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914878 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501-rootfs\") pod \"machine-config-daemon-kswzd\" (UID: \"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\") " pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914881 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-run-multus-certs\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914915 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc-hosts-file\") pod \"node-resolver-lfcbx\" (UID: \"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\") " pod="openshift-dns/node-resolver-lfcbx" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914946 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-var-lib-cni-multus\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914963 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-system-cni-dir\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.914850 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501-rootfs\") pod \"machine-config-daemon-kswzd\" (UID: \"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\") " pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915011 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-var-lib-kubelet\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915012 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-cni-binary-copy\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915082 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-cnibin\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915103 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-var-lib-cni-bin\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915122 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-multus-conf-dir\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915154 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-os-release\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915163 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-cnibin\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915176 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-multus-socket-dir-parent\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915162 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-multus-daemon-config\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915225 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-multus-conf-dir\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915262 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-host-var-lib-cni-bin\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915267 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-multus-socket-dir-parent\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915285 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-os-release\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.915542 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-cni-binary-copy\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.917933 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501-proxy-tls\") pod \"machine-config-daemon-kswzd\" (UID: \"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\") " pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.919336 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.926660 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.926694 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.926702 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.926715 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.926724 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:13Z","lastTransitionTime":"2025-12-04T10:16:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.928702 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.932488 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w667f\" (UniqueName: \"kubernetes.io/projected/b9cb494f-7d4c-45f2-8b9b-e35c42b41c79-kube-api-access-w667f\") pod \"multus-wkbf5\" (UID: \"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\") " pod="openshift-multus/multus-wkbf5" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.937654 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjxng\" (UniqueName: \"kubernetes.io/projected/a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc-kube-api-access-hjxng\") pod \"node-resolver-lfcbx\" (UID: \"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\") " pod="openshift-dns/node-resolver-lfcbx" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.938063 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.939348 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbwhk\" (UniqueName: \"kubernetes.io/projected/7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501-kube-api-access-qbwhk\") pod \"machine-config-daemon-kswzd\" (UID: \"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\") " pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.950415 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.961601 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.969545 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.981021 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:13 crc kubenswrapper[4943]: I1204 10:16:13.992061 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.006888 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.016039 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.016140 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.016182 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.016229 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.016343 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:16:15.016312171 +0000 UTC m=+23.605288039 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.016356 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.016385 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.016411 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:15.016404923 +0000 UTC m=+23.605380791 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.016440 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:15.016424464 +0000 UTC m=+23.605400392 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.016555 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.016572 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.016585 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.016616 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:15.016607828 +0000 UTC m=+23.605583766 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.019091 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.038236 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.038282 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.038293 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.038309 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.038321 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:14Z","lastTransitionTime":"2025-12-04T10:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.042118 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.096867 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.107178 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-wkbf5" Dec 04 10:16:14 crc kubenswrapper[4943]: W1204 10:16:14.119688 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9cb494f_7d4c_45f2_8b9b_e35c42b41c79.slice/crio-86be186ac7914c82170203707e6fe522855e4448d27db1485cea94d062c22b3c WatchSource:0}: Error finding container 86be186ac7914c82170203707e6fe522855e4448d27db1485cea94d062c22b3c: Status 404 returned error can't find the container with id 86be186ac7914c82170203707e6fe522855e4448d27db1485cea94d062c22b3c Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.120792 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-lfcbx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.142013 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.142070 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.142082 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.142105 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.142119 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:14Z","lastTransitionTime":"2025-12-04T10:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.148573 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-9nxvx"] Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.149555 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.151797 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7g4wf"] Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.153352 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.157021 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.157039 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.157957 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.158000 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.158747 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.158827 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.159085 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.159244 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.159414 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 10:16:14 crc kubenswrapper[4943]: W1204 10:16:14.167968 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5a911ae_5fad_4b9a_bfdd_5f2c7f1b33cc.slice/crio-6fd495eaacf367bb808881eef675c5ea421f3da384e8dc3de8cac57d53bf5d9f WatchSource:0}: Error finding container 6fd495eaacf367bb808881eef675c5ea421f3da384e8dc3de8cac57d53bf5d9f: Status 404 returned error can't find the container with id 6fd495eaacf367bb808881eef675c5ea421f3da384e8dc3de8cac57d53bf5d9f Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.174878 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.191255 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.209731 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219312 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-log-socket\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219363 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-kubelet\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219400 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-run-ovn-kubernetes\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219429 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovnkube-config\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219456 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxs29\" (UniqueName: \"kubernetes.io/projected/28b3b2d6-31eb-421d-80f5-71304da27696-kube-api-access-sxs29\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219523 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-ovn\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219568 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/28b3b2d6-31eb-421d-80f5-71304da27696-cnibin\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219610 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-systemd-units\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219635 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-node-log\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219659 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-slash\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219682 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-systemd\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219708 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-openvswitch\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219731 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/28b3b2d6-31eb-421d-80f5-71304da27696-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219780 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219814 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-env-overrides\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219875 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-var-lib-openvswitch\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219898 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-etc-openvswitch\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219921 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wh9m\" (UniqueName: \"kubernetes.io/projected/d0498f58-c99b-4fa8-a404-f7f43199c29f-kube-api-access-7wh9m\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219954 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovnkube-script-lib\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.219999 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/28b3b2d6-31eb-421d-80f5-71304da27696-cni-binary-copy\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.220022 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/28b3b2d6-31eb-421d-80f5-71304da27696-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.220073 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-run-netns\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.220096 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.220130 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/28b3b2d6-31eb-421d-80f5-71304da27696-os-release\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.220177 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-cni-bin\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.220215 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/28b3b2d6-31eb-421d-80f5-71304da27696-system-cni-dir\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.220238 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-cni-netd\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.220258 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovn-node-metrics-cert\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.220266 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.220298 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.220310 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:14 crc kubenswrapper[4943]: E1204 10:16:14.220382 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:15.220360473 +0000 UTC m=+23.809336391 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.229093 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.246661 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.246699 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.246709 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.246726 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.246738 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:14Z","lastTransitionTime":"2025-12-04T10:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.247567 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.265720 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.277300 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.286573 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.296678 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321380 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-cni-netd\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321419 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovn-node-metrics-cert\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321442 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-log-socket\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321479 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-kubelet\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321498 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-run-ovn-kubernetes\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321516 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovnkube-config\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321542 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxs29\" (UniqueName: \"kubernetes.io/projected/28b3b2d6-31eb-421d-80f5-71304da27696-kube-api-access-sxs29\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321565 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-ovn\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321584 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/28b3b2d6-31eb-421d-80f5-71304da27696-cnibin\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321569 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-cni-netd\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321658 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-systemd-units\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321602 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-systemd-units\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321708 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-node-log\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321732 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-slash\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321752 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-systemd\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321773 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/28b3b2d6-31eb-421d-80f5-71304da27696-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321807 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-openvswitch\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321825 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-env-overrides\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321841 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-var-lib-openvswitch\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321857 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-etc-openvswitch\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321876 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wh9m\" (UniqueName: \"kubernetes.io/projected/d0498f58-c99b-4fa8-a404-f7f43199c29f-kube-api-access-7wh9m\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321896 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovnkube-script-lib\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321935 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/28b3b2d6-31eb-421d-80f5-71304da27696-cni-binary-copy\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321955 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/28b3b2d6-31eb-421d-80f5-71304da27696-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321975 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/28b3b2d6-31eb-421d-80f5-71304da27696-os-release\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.321999 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-run-netns\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322022 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322044 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-cni-bin\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322060 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/28b3b2d6-31eb-421d-80f5-71304da27696-system-cni-dir\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322114 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/28b3b2d6-31eb-421d-80f5-71304da27696-system-cni-dir\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322173 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-log-socket\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322273 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-run-ovn-kubernetes\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322299 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-etc-openvswitch\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322322 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-ovn\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322345 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/28b3b2d6-31eb-421d-80f5-71304da27696-cnibin\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322359 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-node-log\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322375 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-slash\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322386 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-systemd\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.322984 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/28b3b2d6-31eb-421d-80f5-71304da27696-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.323008 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-openvswitch\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.323362 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-env-overrides\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.323371 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovnkube-script-lib\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.323388 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-var-lib-openvswitch\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.323418 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.323471 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/28b3b2d6-31eb-421d-80f5-71304da27696-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.323502 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/28b3b2d6-31eb-421d-80f5-71304da27696-os-release\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.323517 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-run-netns\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.323677 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-cni-bin\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.323727 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-kubelet\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.323757 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/28b3b2d6-31eb-421d-80f5-71304da27696-cni-binary-copy\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.323949 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovnkube-config\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.329804 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovn-node-metrics-cert\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.346956 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.353740 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.353784 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.353795 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.353813 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.353824 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:14Z","lastTransitionTime":"2025-12-04T10:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.353893 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxs29\" (UniqueName: \"kubernetes.io/projected/28b3b2d6-31eb-421d-80f5-71304da27696-kube-api-access-sxs29\") pod \"multus-additional-cni-plugins-9nxvx\" (UID: \"28b3b2d6-31eb-421d-80f5-71304da27696\") " pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.356040 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wh9m\" (UniqueName: \"kubernetes.io/projected/d0498f58-c99b-4fa8-a404-f7f43199c29f-kube-api-access-7wh9m\") pod \"ovnkube-node-7g4wf\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.407664 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.456632 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.456662 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.456671 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.456684 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.456694 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:14Z","lastTransitionTime":"2025-12-04T10:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.528146 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.532382 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.560946 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.561018 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.561033 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.561053 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.561068 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:14Z","lastTransitionTime":"2025-12-04T10:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.569151 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.569990 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.571347 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.572050 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.573177 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.573811 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.574577 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.575579 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.576267 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.577298 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.577826 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.579010 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.579556 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.580149 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.581161 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.581762 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.582815 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.583242 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.583866 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.584985 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.585490 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.586524 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.586982 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.588100 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.588561 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.589191 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.590381 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.590888 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.591924 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.592426 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.593429 4943 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.593541 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.595244 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.596172 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.596615 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.598339 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.599015 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.600023 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.600704 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.601844 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.602469 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.603518 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.604394 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.605461 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.605948 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.606901 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.607489 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.608673 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.609235 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.610096 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.610602 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.612216 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.612883 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.613492 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.645498 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.665807 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.665870 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.665888 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.665909 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.665920 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:14Z","lastTransitionTime":"2025-12-04T10:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.672178 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.713116 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.724436 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.727254 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.727302 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d7dbc4b292bd4ff5b1b50087184766e52a698412f33d771c2db9770191825f89"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.730219 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.730332 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.732896 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"e9bbac1f390cd447ef4c261ccdc5cf69edafa1c4c5d23cae22c29d3289bf8d17"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.735078 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lfcbx" event={"ID":"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc","Type":"ContainerStarted","Data":"6fd495eaacf367bb808881eef675c5ea421f3da384e8dc3de8cac57d53bf5d9f"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.735910 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.737650 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wkbf5" event={"ID":"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79","Type":"ContainerStarted","Data":"babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.737711 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wkbf5" event={"ID":"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79","Type":"ContainerStarted","Data":"86be186ac7914c82170203707e6fe522855e4448d27db1485cea94d062c22b3c"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.746905 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4aaf76b39270c97deee766a204b04fae5be13b5ee2f290d035c816e0daa1681d"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.751778 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" event={"ID":"28b3b2d6-31eb-421d-80f5-71304da27696","Type":"ContainerStarted","Data":"4cc01e5d2cd5aeed69835a2a88faefc38f652da38a43e24a8652507277e8cc7a"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.754240 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerStarted","Data":"9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.754285 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerStarted","Data":"f08291677c0e6c6a4a7806c51ab40368e0906f33c7d41b5f4edd3754bfa8c2fd"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.757995 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.786069 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.787105 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.787138 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.787153 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.787265 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.787286 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:14Z","lastTransitionTime":"2025-12-04T10:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.803793 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.822808 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.845902 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.873959 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.889704 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.889748 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.889760 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.889776 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.889786 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:14Z","lastTransitionTime":"2025-12-04T10:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.898879 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.915695 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.937671 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.952684 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.965168 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.982941 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:14Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.991768 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.991802 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.991812 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.991826 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:14 crc kubenswrapper[4943]: I1204 10:16:14.991835 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:14Z","lastTransitionTime":"2025-12-04T10:16:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.003712 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.018353 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.028841 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.028997 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.029040 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:16:17.029014007 +0000 UTC m=+25.617989875 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.029120 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.029139 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.029150 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.029144 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.029184 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.029218 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:17.029182621 +0000 UTC m=+25.618158489 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.029284 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.029331 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:17.029322784 +0000 UTC m=+25.618298752 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.029370 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.029458 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:17.029440477 +0000 UTC m=+25.618416345 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.036233 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.054919 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.070951 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.082877 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.094009 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.094045 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.094053 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.094069 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.094079 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:15Z","lastTransitionTime":"2025-12-04T10:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.099483 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.196515 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.196558 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.196566 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.196581 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.196592 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:15Z","lastTransitionTime":"2025-12-04T10:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.230259 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.230458 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.230483 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.230497 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.230561 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:17.230543658 +0000 UTC m=+25.819519526 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.301553 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.301607 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.301619 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.301636 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.301648 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:15Z","lastTransitionTime":"2025-12-04T10:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.404504 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.404554 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.404566 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.404586 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.404598 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:15Z","lastTransitionTime":"2025-12-04T10:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.506986 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.507032 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.507042 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.507057 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.507067 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:15Z","lastTransitionTime":"2025-12-04T10:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.563893 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.563948 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.563905 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.564028 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.564316 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:15 crc kubenswrapper[4943]: E1204 10:16:15.564178 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.609265 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.609310 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.609319 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.609337 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.609348 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:15Z","lastTransitionTime":"2025-12-04T10:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.711512 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.711551 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.711565 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.711583 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.711603 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:15Z","lastTransitionTime":"2025-12-04T10:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.757643 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966" exitCode=0 Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.757743 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.759100 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-lfcbx" event={"ID":"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc","Type":"ContainerStarted","Data":"dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.761125 4943 generic.go:334] "Generic (PLEG): container finished" podID="28b3b2d6-31eb-421d-80f5-71304da27696" containerID="e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8" exitCode=0 Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.761211 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" event={"ID":"28b3b2d6-31eb-421d-80f5-71304da27696","Type":"ContainerDied","Data":"e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.767237 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerStarted","Data":"f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.786153 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.804152 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.822083 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.822807 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.822858 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.822873 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.822896 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.822906 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:15Z","lastTransitionTime":"2025-12-04T10:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.839306 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.851740 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.861853 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.885564 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.898745 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.922562 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.926645 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.926687 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.926698 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.926714 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.926724 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:15Z","lastTransitionTime":"2025-12-04T10:16:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.939919 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.953574 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.967121 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.980616 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:15 crc kubenswrapper[4943]: I1204 10:16:15.997886 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.016086 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.028828 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.028879 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.028893 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.028910 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.028921 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:16Z","lastTransitionTime":"2025-12-04T10:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.038725 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.056117 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.069216 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.087749 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.102084 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.120675 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.131117 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.131160 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.131170 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.131185 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.131219 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:16Z","lastTransitionTime":"2025-12-04T10:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.138860 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.159027 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.177874 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.234380 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.234796 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.234811 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.234830 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.234842 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:16Z","lastTransitionTime":"2025-12-04T10:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.417383 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.417467 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.417478 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.417496 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.417512 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:16Z","lastTransitionTime":"2025-12-04T10:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.520555 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.520611 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.520625 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.520643 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.520652 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:16Z","lastTransitionTime":"2025-12-04T10:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.632391 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.632434 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.632445 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.632460 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.632469 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:16Z","lastTransitionTime":"2025-12-04T10:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.735078 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.735115 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.735124 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.735137 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.735146 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:16Z","lastTransitionTime":"2025-12-04T10:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.781426 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" event={"ID":"28b3b2d6-31eb-421d-80f5-71304da27696","Type":"ContainerStarted","Data":"025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.785797 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.785833 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.785842 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.785855 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.837890 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.837918 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.837926 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.837940 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.837949 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:16Z","lastTransitionTime":"2025-12-04T10:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.838744 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.854545 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.869355 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.882086 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.894774 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.912580 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.926131 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.938528 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.939794 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.939833 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.939842 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.939856 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.939916 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:16Z","lastTransitionTime":"2025-12-04T10:16:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.950955 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.972547 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:16 crc kubenswrapper[4943]: I1204 10:16:16.989806 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.001735 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:16Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.043926 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.043966 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.043975 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.044010 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.044025 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:17Z","lastTransitionTime":"2025-12-04T10:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.120593 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.120739 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.120922 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:16:21.120825819 +0000 UTC m=+29.709801687 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.120929 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.120964 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.120991 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.121004 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:21.120997753 +0000 UTC m=+29.709973621 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.121098 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.121119 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.121135 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.121186 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:21.121172897 +0000 UTC m=+29.710148765 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.121176 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.121581 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:21.121518346 +0000 UTC m=+29.710494274 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.147155 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.147194 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.147226 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.147243 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.147256 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:17Z","lastTransitionTime":"2025-12-04T10:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.250223 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.250276 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.250289 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.250310 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.250322 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:17Z","lastTransitionTime":"2025-12-04T10:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.323133 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.323374 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.323670 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.323689 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.323737 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:21.323723113 +0000 UTC m=+29.912698981 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.353108 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.353155 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.353166 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.353181 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.353191 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:17Z","lastTransitionTime":"2025-12-04T10:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.455610 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.455658 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.455668 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.455683 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.455694 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:17Z","lastTransitionTime":"2025-12-04T10:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.558653 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.558694 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.558703 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.558718 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.558728 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:17Z","lastTransitionTime":"2025-12-04T10:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.564072 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.564073 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.564221 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.564263 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.564442 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:17 crc kubenswrapper[4943]: E1204 10:16:17.564619 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.663108 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.663543 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.663707 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.663805 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.663869 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:17Z","lastTransitionTime":"2025-12-04T10:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.767852 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.767908 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.767921 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.767942 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.767957 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:17Z","lastTransitionTime":"2025-12-04T10:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.797836 4943 generic.go:334] "Generic (PLEG): container finished" podID="28b3b2d6-31eb-421d-80f5-71304da27696" containerID="025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032" exitCode=0 Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.798072 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" event={"ID":"28b3b2d6-31eb-421d-80f5-71304da27696","Type":"ContainerDied","Data":"025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032"} Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.808521 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98"} Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.808749 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744"} Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.812163 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:17Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.825590 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:17Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.847022 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:17Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.864675 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:17Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.871527 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.871567 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.871579 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.871598 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.871612 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:17Z","lastTransitionTime":"2025-12-04T10:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.881862 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:17Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.894693 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:17Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.909433 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:17Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.920485 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:17Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.930230 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:17Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.940627 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:17Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.954539 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:17Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.971654 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:17Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.973235 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.973263 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.973271 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.973285 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:17 crc kubenswrapper[4943]: I1204 10:16:17.973294 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:17Z","lastTransitionTime":"2025-12-04T10:16:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.077541 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.077591 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.077604 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.077621 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.077666 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:18Z","lastTransitionTime":"2025-12-04T10:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.180461 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.180502 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.180512 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.180531 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.180543 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:18Z","lastTransitionTime":"2025-12-04T10:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.283611 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.283656 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.283669 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.283685 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.283696 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:18Z","lastTransitionTime":"2025-12-04T10:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.386900 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.386947 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.386956 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.387164 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.387174 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:18Z","lastTransitionTime":"2025-12-04T10:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.489743 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.489808 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.489826 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.489852 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.489868 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:18Z","lastTransitionTime":"2025-12-04T10:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.593061 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.593097 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.593104 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.593118 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.593129 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:18Z","lastTransitionTime":"2025-12-04T10:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.696218 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.696263 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.696303 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.696325 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.696337 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:18Z","lastTransitionTime":"2025-12-04T10:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.799463 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.799508 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.799518 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.799533 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.799542 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:18Z","lastTransitionTime":"2025-12-04T10:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.812735 4943 generic.go:334] "Generic (PLEG): container finished" podID="28b3b2d6-31eb-421d-80f5-71304da27696" containerID="0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223" exitCode=0 Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.812777 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" event={"ID":"28b3b2d6-31eb-421d-80f5-71304da27696","Type":"ContainerDied","Data":"0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223"} Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.814821 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987"} Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.850989 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.865785 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.881392 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.895762 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.901809 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.901858 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.901871 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.901889 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.901903 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:18Z","lastTransitionTime":"2025-12-04T10:16:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.911009 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.922002 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.935100 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.953894 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.968535 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.985051 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:18 crc kubenswrapper[4943]: I1204 10:16:18.998596 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.003689 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.003828 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.003888 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.003950 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.004005 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:19Z","lastTransitionTime":"2025-12-04T10:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.023867 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.117164 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.117223 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.117236 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.117254 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.117271 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:19Z","lastTransitionTime":"2025-12-04T10:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.117681 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.145532 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.163031 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.175895 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.190451 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.202920 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.219560 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.219605 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.219616 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.219632 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.219643 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:19Z","lastTransitionTime":"2025-12-04T10:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.219836 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.230432 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.238549 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.249569 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.259858 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.270617 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.322561 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.322601 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.322611 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.322626 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.322637 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:19Z","lastTransitionTime":"2025-12-04T10:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.425416 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.425478 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.425491 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.425522 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.425535 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:19Z","lastTransitionTime":"2025-12-04T10:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.527990 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.528062 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.528076 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.528113 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.528127 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:19Z","lastTransitionTime":"2025-12-04T10:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.563673 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.563712 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.563683 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:19 crc kubenswrapper[4943]: E1204 10:16:19.563850 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:19 crc kubenswrapper[4943]: E1204 10:16:19.564001 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:19 crc kubenswrapper[4943]: E1204 10:16:19.564095 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.630554 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.630592 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.630602 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.630622 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.630633 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:19Z","lastTransitionTime":"2025-12-04T10:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.733690 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.733751 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.733761 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.733776 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.733786 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:19Z","lastTransitionTime":"2025-12-04T10:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.821990 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284"} Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.824956 4943 generic.go:334] "Generic (PLEG): container finished" podID="28b3b2d6-31eb-421d-80f5-71304da27696" containerID="00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d" exitCode=0 Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.825095 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" event={"ID":"28b3b2d6-31eb-421d-80f5-71304da27696","Type":"ContainerDied","Data":"00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d"} Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.836345 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.836394 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.836405 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.836424 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.836437 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:19Z","lastTransitionTime":"2025-12-04T10:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.839404 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.856788 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.868338 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.881828 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.895358 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.913649 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.925650 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.934066 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.938945 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.939002 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.939014 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.939031 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.939044 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:19Z","lastTransitionTime":"2025-12-04T10:16:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.946724 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.960608 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.972169 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:19 crc kubenswrapper[4943]: I1204 10:16:19.982361 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:19Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.040880 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.040928 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.040938 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.040956 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.040970 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:20Z","lastTransitionTime":"2025-12-04T10:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.143867 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.143914 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.143925 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.143939 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.143949 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:20Z","lastTransitionTime":"2025-12-04T10:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.246278 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.246341 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.246351 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.246366 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.246582 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:20Z","lastTransitionTime":"2025-12-04T10:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.349069 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.349106 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.349117 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.349134 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.349147 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:20Z","lastTransitionTime":"2025-12-04T10:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.451930 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.451974 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.451985 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.452001 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.452013 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:20Z","lastTransitionTime":"2025-12-04T10:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.509075 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-5njp4"] Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.509901 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5njp4" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.512768 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.512957 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.512875 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.513072 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.530934 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.550461 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.553405 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.555124 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.555176 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.555193 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.555242 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.555259 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:20Z","lastTransitionTime":"2025-12-04T10:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.559193 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.562815 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eeee576e-fae5-42f5-9248-0864812c01aa-host\") pod \"node-ca-5njp4\" (UID: \"eeee576e-fae5-42f5-9248-0864812c01aa\") " pod="openshift-image-registry/node-ca-5njp4" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.562895 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/eeee576e-fae5-42f5-9248-0864812c01aa-serviceca\") pod \"node-ca-5njp4\" (UID: \"eeee576e-fae5-42f5-9248-0864812c01aa\") " pod="openshift-image-registry/node-ca-5njp4" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.562977 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p5w8\" (UniqueName: \"kubernetes.io/projected/eeee576e-fae5-42f5-9248-0864812c01aa-kube-api-access-7p5w8\") pod \"node-ca-5njp4\" (UID: \"eeee576e-fae5-42f5-9248-0864812c01aa\") " pod="openshift-image-registry/node-ca-5njp4" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.565243 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.568134 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.580124 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.592445 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.602647 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.612456 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.625679 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.640385 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.653052 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.657045 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.657927 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.657987 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.658019 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.658033 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:20Z","lastTransitionTime":"2025-12-04T10:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.664086 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eeee576e-fae5-42f5-9248-0864812c01aa-host\") pod \"node-ca-5njp4\" (UID: \"eeee576e-fae5-42f5-9248-0864812c01aa\") " pod="openshift-image-registry/node-ca-5njp4" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.664235 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eeee576e-fae5-42f5-9248-0864812c01aa-host\") pod \"node-ca-5njp4\" (UID: \"eeee576e-fae5-42f5-9248-0864812c01aa\") " pod="openshift-image-registry/node-ca-5njp4" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.664270 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/eeee576e-fae5-42f5-9248-0864812c01aa-serviceca\") pod \"node-ca-5njp4\" (UID: \"eeee576e-fae5-42f5-9248-0864812c01aa\") " pod="openshift-image-registry/node-ca-5njp4" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.664404 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p5w8\" (UniqueName: \"kubernetes.io/projected/eeee576e-fae5-42f5-9248-0864812c01aa-kube-api-access-7p5w8\") pod \"node-ca-5njp4\" (UID: \"eeee576e-fae5-42f5-9248-0864812c01aa\") " pod="openshift-image-registry/node-ca-5njp4" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.669573 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.672702 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/eeee576e-fae5-42f5-9248-0864812c01aa-serviceca\") pod \"node-ca-5njp4\" (UID: \"eeee576e-fae5-42f5-9248-0864812c01aa\") " pod="openshift-image-registry/node-ca-5njp4" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.683103 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p5w8\" (UniqueName: \"kubernetes.io/projected/eeee576e-fae5-42f5-9248-0864812c01aa-kube-api-access-7p5w8\") pod \"node-ca-5njp4\" (UID: \"eeee576e-fae5-42f5-9248-0864812c01aa\") " pod="openshift-image-registry/node-ca-5njp4" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.683720 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.706191 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.719250 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.733531 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.745661 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.761473 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.761565 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.761582 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.761600 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.761610 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:20Z","lastTransitionTime":"2025-12-04T10:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.762699 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.775425 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.788891 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.801153 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.816268 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.826665 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.826705 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5njp4" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.831937 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" event={"ID":"28b3b2d6-31eb-421d-80f5-71304da27696","Type":"ContainerStarted","Data":"dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c"} Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.838733 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: E1204 10:16:20.839467 4943 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:16:20 crc kubenswrapper[4943]: W1204 10:16:20.842995 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeeee576e_fae5_42f5_9248_0864812c01aa.slice/crio-f50b0edbb44758e196e14d5ad44afec9bc787330d09d73e851c6d18db49454f6 WatchSource:0}: Error finding container f50b0edbb44758e196e14d5ad44afec9bc787330d09d73e851c6d18db49454f6: Status 404 returned error can't find the container with id f50b0edbb44758e196e14d5ad44afec9bc787330d09d73e851c6d18db49454f6 Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.851823 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.865340 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.865381 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.865391 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.865408 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.865418 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:20Z","lastTransitionTime":"2025-12-04T10:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.869428 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.882647 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.904560 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.917014 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.928665 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.941838 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.967789 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.967851 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.967860 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.967876 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.967885 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:20Z","lastTransitionTime":"2025-12-04T10:16:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.977732 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:20 crc kubenswrapper[4943]: I1204 10:16:20.993039 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:20Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.005110 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:21Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.018575 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:21Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.033095 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:21Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.047931 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:21Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.061004 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:21Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.070751 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.070786 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.070798 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.070813 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.070824 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:21Z","lastTransitionTime":"2025-12-04T10:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.075286 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:21Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.098989 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:21Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.112091 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:21Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.127727 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:21Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.169626 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.169779 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.169818 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.169851 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.169947 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:16:29.169914451 +0000 UTC m=+37.758890319 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.169987 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.170042 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.170041 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.170121 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.170138 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.170096 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:29.170068004 +0000 UTC m=+37.759043872 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.170221 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:29.170195387 +0000 UTC m=+37.759171415 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.170238 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:29.170228828 +0000 UTC m=+37.759204916 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.174133 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.174224 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.174242 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.174267 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.174283 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:21Z","lastTransitionTime":"2025-12-04T10:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.276767 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.276815 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.276830 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.276851 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.276864 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:21Z","lastTransitionTime":"2025-12-04T10:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.371732 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.371934 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.371959 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.371973 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.372054 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:29.372036636 +0000 UTC m=+37.961012504 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.379278 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.379345 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.379361 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.379389 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.379405 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:21Z","lastTransitionTime":"2025-12-04T10:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.483559 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.483611 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.483622 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.483641 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.483655 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:21Z","lastTransitionTime":"2025-12-04T10:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.564454 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.564496 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.564496 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.564664 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.564787 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:21 crc kubenswrapper[4943]: E1204 10:16:21.564889 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.586695 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.586773 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.586800 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.586822 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.586834 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:21Z","lastTransitionTime":"2025-12-04T10:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.689327 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.689381 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.689394 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.689410 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.689432 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:21Z","lastTransitionTime":"2025-12-04T10:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.792497 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.792553 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.792584 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.792609 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.792626 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:21Z","lastTransitionTime":"2025-12-04T10:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.895088 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.895155 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.895166 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.895193 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.895259 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:21Z","lastTransitionTime":"2025-12-04T10:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.998450 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.998504 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.998516 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.998533 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:21 crc kubenswrapper[4943]: I1204 10:16:21.998544 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:21Z","lastTransitionTime":"2025-12-04T10:16:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.101316 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.101371 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.101385 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.101406 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.101421 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:22Z","lastTransitionTime":"2025-12-04T10:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.124100 4943 generic.go:334] "Generic (PLEG): container finished" podID="28b3b2d6-31eb-421d-80f5-71304da27696" containerID="dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c" exitCode=0 Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.124399 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" event={"ID":"28b3b2d6-31eb-421d-80f5-71304da27696","Type":"ContainerDied","Data":"dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c"} Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.126385 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5njp4" event={"ID":"eeee576e-fae5-42f5-9248-0864812c01aa","Type":"ContainerStarted","Data":"d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3"} Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.126437 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5njp4" event={"ID":"eeee576e-fae5-42f5-9248-0864812c01aa","Type":"ContainerStarted","Data":"f50b0edbb44758e196e14d5ad44afec9bc787330d09d73e851c6d18db49454f6"} Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.143179 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.158846 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.174304 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.197772 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.210921 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.210991 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.211004 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.211042 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.211056 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:22Z","lastTransitionTime":"2025-12-04T10:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.223511 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.238519 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.252319 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.270182 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.290687 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.304697 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.314477 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.314545 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.314560 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.314580 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.314592 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:22Z","lastTransitionTime":"2025-12-04T10:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.317325 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.332439 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.348415 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.364316 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.381997 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.394800 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.417395 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.417466 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.417481 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.417507 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.417523 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:22Z","lastTransitionTime":"2025-12-04T10:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.417678 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.435837 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.450335 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.462422 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.473365 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.486352 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.498685 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.510583 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.519939 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.520011 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.520023 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.520040 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.520051 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:22Z","lastTransitionTime":"2025-12-04T10:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.521389 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.536774 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.550271 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.564100 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.578863 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.590147 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.607469 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.617822 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.621959 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.621989 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.621999 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.622015 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.622027 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:22Z","lastTransitionTime":"2025-12-04T10:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.628556 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.643565 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.663588 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.674634 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.688884 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.713067 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.723988 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.724037 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.724046 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.724061 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.724072 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:22Z","lastTransitionTime":"2025-12-04T10:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.727607 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.740191 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.750963 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.764497 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:22Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.826916 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.826971 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.826985 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.827007 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.827050 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:22Z","lastTransitionTime":"2025-12-04T10:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.928889 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.928943 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.928952 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.928965 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:22 crc kubenswrapper[4943]: I1204 10:16:22.928974 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:22Z","lastTransitionTime":"2025-12-04T10:16:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.031319 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.031363 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.031374 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.031389 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.031400 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.132813 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.132849 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.132862 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.132878 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.132890 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.134183 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89"} Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.134616 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.134634 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.140651 4943 generic.go:334] "Generic (PLEG): container finished" podID="28b3b2d6-31eb-421d-80f5-71304da27696" containerID="29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667" exitCode=0 Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.140701 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" event={"ID":"28b3b2d6-31eb-421d-80f5-71304da27696","Type":"ContainerDied","Data":"29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667"} Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.146926 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.159569 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.173661 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.181602 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.200896 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.216552 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.235904 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.238145 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.238171 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.238246 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.238265 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.238277 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.248442 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.263454 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.274732 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.284984 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.294609 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.303196 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.311762 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.322479 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.332525 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.340943 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.340982 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.340994 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.341010 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.341023 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.348572 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.358103 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.368646 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.378986 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.390060 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.400525 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.409789 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.419501 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.432722 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.443828 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.443864 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.443874 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.443887 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.443897 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.445888 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.457659 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.471542 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.482223 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.546543 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.546586 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.546596 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.546615 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.546642 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.563851 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.563848 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.563922 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:23 crc kubenswrapper[4943]: E1204 10:16:23.564011 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:23 crc kubenswrapper[4943]: E1204 10:16:23.564219 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:23 crc kubenswrapper[4943]: E1204 10:16:23.564360 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.648688 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.648736 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.648747 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.648766 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.648776 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.766652 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.766685 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.766693 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.766708 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.766718 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.868889 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.869447 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.869529 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.869602 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.869667 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.898285 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.898332 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.898343 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.898361 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.898375 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: E1204 10:16:23.912492 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.915750 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.915808 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.915821 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.915840 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.915853 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: E1204 10:16:23.927547 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.931080 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.931123 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.931133 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.931148 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.931158 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: E1204 10:16:23.944513 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.948942 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.949184 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.949296 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.949388 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.949458 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: E1204 10:16:23.965811 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.974180 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.974236 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.974245 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.974260 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.974269 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:23 crc kubenswrapper[4943]: E1204 10:16:23.986586 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:23Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:23 crc kubenswrapper[4943]: E1204 10:16:23.986716 4943 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.988340 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.988373 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.988381 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.988395 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:23 crc kubenswrapper[4943]: I1204 10:16:23.988404 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:23Z","lastTransitionTime":"2025-12-04T10:16:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.090901 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.090982 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.090994 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.091011 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.091022 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:24Z","lastTransitionTime":"2025-12-04T10:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.206468 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.206516 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.206528 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.206547 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.206561 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:24Z","lastTransitionTime":"2025-12-04T10:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.226394 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" event={"ID":"28b3b2d6-31eb-421d-80f5-71304da27696","Type":"ContainerStarted","Data":"3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c"} Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.226962 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.239040 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.254574 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.259518 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.269014 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.284381 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.311722 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.313672 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.313711 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.313720 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.313733 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.313743 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:24Z","lastTransitionTime":"2025-12-04T10:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.329897 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.348497 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.372463 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.387836 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.402044 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.412740 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.416175 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.416217 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.416226 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.416239 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.416247 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:24Z","lastTransitionTime":"2025-12-04T10:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.423957 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.436008 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.447324 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.459579 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.471443 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.488318 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.507546 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.545091 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.545138 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.545150 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.545166 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.545176 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:24Z","lastTransitionTime":"2025-12-04T10:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.549507 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.560131 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.570023 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.585458 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.603349 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.617874 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.627762 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.644190 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.647392 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.647420 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.647429 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.647443 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.647452 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:24Z","lastTransitionTime":"2025-12-04T10:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.656536 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.669529 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:24Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.749688 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.749739 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.749778 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.749795 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.749805 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:24Z","lastTransitionTime":"2025-12-04T10:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.853045 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.853132 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.853147 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.853191 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.853224 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:24Z","lastTransitionTime":"2025-12-04T10:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.955907 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.955955 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.955966 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.955982 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:24 crc kubenswrapper[4943]: I1204 10:16:24.955997 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:24Z","lastTransitionTime":"2025-12-04T10:16:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.058815 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.058869 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.058881 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.058903 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.058916 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:25Z","lastTransitionTime":"2025-12-04T10:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.161683 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.161724 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.161732 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.161747 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.161757 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:25Z","lastTransitionTime":"2025-12-04T10:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.231293 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/0.log" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.233679 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89" exitCode=1 Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.233769 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89"} Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.234612 4943 scope.go:117] "RemoveContainer" containerID="8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.245859 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.261787 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.265604 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.265650 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.265661 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.265682 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.265695 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:25Z","lastTransitionTime":"2025-12-04T10:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.276266 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.291032 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.311473 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:25Z\\\",\\\"message\\\":\\\"0:16:25.017976 6145 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 10:16:25.018126 6145 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 10:16:25.018310 6145 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:25.018453 6145 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018512 6145 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018817 6145 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 10:16:25.018867 6145 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 10:16:25.018886 6145 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 10:16:25.018891 6145 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 10:16:25.018920 6145 factory.go:656] Stopping watch factory\\\\nI1204 10:16:25.018935 6145 ovnkube.go:599] Stopped ovnkube\\\\nI1204 10:16:25.018949 6145 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 10:16:25.018966 6145 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 10:16:25.018961 6145 handler.go:208] Removed *v1.Pod event ha\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.323563 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.336920 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.352193 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.366499 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.367754 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.367821 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.367835 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.367854 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.367869 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:25Z","lastTransitionTime":"2025-12-04T10:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.381377 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.393511 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.404880 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.419303 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.432529 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.470838 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.470897 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.470911 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.470929 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.470943 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:25Z","lastTransitionTime":"2025-12-04T10:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.564253 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.564299 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.564253 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:25 crc kubenswrapper[4943]: E1204 10:16:25.564426 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:25 crc kubenswrapper[4943]: E1204 10:16:25.564534 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:25 crc kubenswrapper[4943]: E1204 10:16:25.564796 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.576082 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.576165 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.576179 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.576227 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.576245 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:25Z","lastTransitionTime":"2025-12-04T10:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.679074 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.679127 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.679138 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.679154 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.679164 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:25Z","lastTransitionTime":"2025-12-04T10:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.781718 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.781767 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.781778 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.782144 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.782812 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:25Z","lastTransitionTime":"2025-12-04T10:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.885583 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.885668 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.885681 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.885706 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.885723 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:25Z","lastTransitionTime":"2025-12-04T10:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.988041 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.988098 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.988110 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.988130 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:25 crc kubenswrapper[4943]: I1204 10:16:25.988143 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:25Z","lastTransitionTime":"2025-12-04T10:16:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.090643 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.090692 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.090703 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.090720 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.090732 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:26Z","lastTransitionTime":"2025-12-04T10:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.192814 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.192871 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.192890 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.192909 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.192922 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:26Z","lastTransitionTime":"2025-12-04T10:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.239962 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/0.log" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.242387 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1"} Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.242926 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.258097 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.270632 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.281069 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.293083 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.294802 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.294843 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.294856 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.294874 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.294886 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:26Z","lastTransitionTime":"2025-12-04T10:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.305135 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.317279 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.329567 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.344338 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.360635 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.376578 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.389728 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.397896 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.397973 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.398003 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.398029 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.398045 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:26Z","lastTransitionTime":"2025-12-04T10:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.402780 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.421463 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:25Z\\\",\\\"message\\\":\\\"0:16:25.017976 6145 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 10:16:25.018126 6145 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 10:16:25.018310 6145 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:25.018453 6145 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018512 6145 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018817 6145 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 10:16:25.018867 6145 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 10:16:25.018886 6145 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 10:16:25.018891 6145 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 10:16:25.018920 6145 factory.go:656] Stopping watch factory\\\\nI1204 10:16:25.018935 6145 ovnkube.go:599] Stopped ovnkube\\\\nI1204 10:16:25.018949 6145 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 10:16:25.018966 6145 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 10:16:25.018961 6145 handler.go:208] Removed *v1.Pod event ha\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.433220 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:26Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.501954 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.502006 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.502026 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.502045 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.502056 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:26Z","lastTransitionTime":"2025-12-04T10:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.604263 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.604340 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.604358 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.604381 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.604401 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:26Z","lastTransitionTime":"2025-12-04T10:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.709454 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.709535 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.709562 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.709596 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.709617 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:26Z","lastTransitionTime":"2025-12-04T10:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.812729 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.812780 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.812795 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.812815 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.812830 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:26Z","lastTransitionTime":"2025-12-04T10:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.914767 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.915053 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.915175 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.915344 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:26 crc kubenswrapper[4943]: I1204 10:16:26.915473 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:26Z","lastTransitionTime":"2025-12-04T10:16:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.018270 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.018341 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.018355 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.018375 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.018388 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:27Z","lastTransitionTime":"2025-12-04T10:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.120858 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.120920 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.120931 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.120949 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.120961 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:27Z","lastTransitionTime":"2025-12-04T10:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.223406 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.223454 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.223465 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.223482 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.223493 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:27Z","lastTransitionTime":"2025-12-04T10:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.326293 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.326341 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.326352 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.326367 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.326379 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:27Z","lastTransitionTime":"2025-12-04T10:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.433972 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.434071 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.434092 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.434119 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.434144 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:27Z","lastTransitionTime":"2025-12-04T10:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.537568 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.537627 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.537644 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.537677 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.537695 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:27Z","lastTransitionTime":"2025-12-04T10:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.563716 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.563737 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.563816 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:27 crc kubenswrapper[4943]: E1204 10:16:27.564423 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:27 crc kubenswrapper[4943]: E1204 10:16:27.564604 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:27 crc kubenswrapper[4943]: E1204 10:16:27.564764 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.590611 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd"] Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.591247 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.593837 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.594555 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.606274 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.620997 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.633900 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.639987 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.640062 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.640082 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.640113 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.640134 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:27Z","lastTransitionTime":"2025-12-04T10:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.652184 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.665773 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.667292 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b13e11ac-5073-4735-bc48-196c6ebeaeaa-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-v6hxd\" (UID: \"b13e11ac-5073-4735-bc48-196c6ebeaeaa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.667367 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtxrq\" (UniqueName: \"kubernetes.io/projected/b13e11ac-5073-4735-bc48-196c6ebeaeaa-kube-api-access-dtxrq\") pod \"ovnkube-control-plane-749d76644c-v6hxd\" (UID: \"b13e11ac-5073-4735-bc48-196c6ebeaeaa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.667406 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b13e11ac-5073-4735-bc48-196c6ebeaeaa-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-v6hxd\" (UID: \"b13e11ac-5073-4735-bc48-196c6ebeaeaa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.667612 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b13e11ac-5073-4735-bc48-196c6ebeaeaa-env-overrides\") pod \"ovnkube-control-plane-749d76644c-v6hxd\" (UID: \"b13e11ac-5073-4735-bc48-196c6ebeaeaa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.678093 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.692040 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.710610 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:25Z\\\",\\\"message\\\":\\\"0:16:25.017976 6145 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 10:16:25.018126 6145 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 10:16:25.018310 6145 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:25.018453 6145 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018512 6145 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018817 6145 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 10:16:25.018867 6145 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 10:16:25.018886 6145 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 10:16:25.018891 6145 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 10:16:25.018920 6145 factory.go:656] Stopping watch factory\\\\nI1204 10:16:25.018935 6145 ovnkube.go:599] Stopped ovnkube\\\\nI1204 10:16:25.018949 6145 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 10:16:25.018966 6145 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 10:16:25.018961 6145 handler.go:208] Removed *v1.Pod event ha\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.719300 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.729688 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.741502 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.742617 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.742672 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.742686 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.742704 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.742715 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:27Z","lastTransitionTime":"2025-12-04T10:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.755823 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.768523 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b13e11ac-5073-4735-bc48-196c6ebeaeaa-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-v6hxd\" (UID: \"b13e11ac-5073-4735-bc48-196c6ebeaeaa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.768580 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtxrq\" (UniqueName: \"kubernetes.io/projected/b13e11ac-5073-4735-bc48-196c6ebeaeaa-kube-api-access-dtxrq\") pod \"ovnkube-control-plane-749d76644c-v6hxd\" (UID: \"b13e11ac-5073-4735-bc48-196c6ebeaeaa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.768607 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b13e11ac-5073-4735-bc48-196c6ebeaeaa-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-v6hxd\" (UID: \"b13e11ac-5073-4735-bc48-196c6ebeaeaa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.768626 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b13e11ac-5073-4735-bc48-196c6ebeaeaa-env-overrides\") pod \"ovnkube-control-plane-749d76644c-v6hxd\" (UID: \"b13e11ac-5073-4735-bc48-196c6ebeaeaa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.769245 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b13e11ac-5073-4735-bc48-196c6ebeaeaa-env-overrides\") pod \"ovnkube-control-plane-749d76644c-v6hxd\" (UID: \"b13e11ac-5073-4735-bc48-196c6ebeaeaa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.769463 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b13e11ac-5073-4735-bc48-196c6ebeaeaa-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-v6hxd\" (UID: \"b13e11ac-5073-4735-bc48-196c6ebeaeaa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.770685 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.775466 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b13e11ac-5073-4735-bc48-196c6ebeaeaa-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-v6hxd\" (UID: \"b13e11ac-5073-4735-bc48-196c6ebeaeaa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.782549 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.784667 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtxrq\" (UniqueName: \"kubernetes.io/projected/b13e11ac-5073-4735-bc48-196c6ebeaeaa-kube-api-access-dtxrq\") pod \"ovnkube-control-plane-749d76644c-v6hxd\" (UID: \"b13e11ac-5073-4735-bc48-196c6ebeaeaa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.793254 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:27Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.845002 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.845050 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.845058 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.845073 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.845082 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:27Z","lastTransitionTime":"2025-12-04T10:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.903990 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" Dec 04 10:16:27 crc kubenswrapper[4943]: W1204 10:16:27.915910 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb13e11ac_5073_4735_bc48_196c6ebeaeaa.slice/crio-080b5a4146e8bc269eb94feed65f28dc3fc1f5f791985b307905ff25288322ef WatchSource:0}: Error finding container 080b5a4146e8bc269eb94feed65f28dc3fc1f5f791985b307905ff25288322ef: Status 404 returned error can't find the container with id 080b5a4146e8bc269eb94feed65f28dc3fc1f5f791985b307905ff25288322ef Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.948303 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.948647 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.948733 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.948835 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:27 crc kubenswrapper[4943]: I1204 10:16:27.948919 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:27Z","lastTransitionTime":"2025-12-04T10:16:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.051507 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.051547 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.051556 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.051570 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.051579 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:28Z","lastTransitionTime":"2025-12-04T10:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.154604 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.154659 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.154677 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.154695 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.154711 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:28Z","lastTransitionTime":"2025-12-04T10:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.248958 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" event={"ID":"b13e11ac-5073-4735-bc48-196c6ebeaeaa","Type":"ContainerStarted","Data":"7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.249020 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" event={"ID":"b13e11ac-5073-4735-bc48-196c6ebeaeaa","Type":"ContainerStarted","Data":"c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.249036 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" event={"ID":"b13e11ac-5073-4735-bc48-196c6ebeaeaa","Type":"ContainerStarted","Data":"080b5a4146e8bc269eb94feed65f28dc3fc1f5f791985b307905ff25288322ef"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.251225 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/1.log" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.251785 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/0.log" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.254374 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1" exitCode=1 Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.254408 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.254440 4943 scope.go:117] "RemoveContainer" containerID="8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.255231 4943 scope.go:117] "RemoveContainer" containerID="da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1" Dec 04 10:16:28 crc kubenswrapper[4943]: E1204 10:16:28.255403 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.256731 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.256756 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.256767 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.256782 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.256794 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:28Z","lastTransitionTime":"2025-12-04T10:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.266507 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.284224 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.305255 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.322169 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-hw5pj"] Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.322077 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:25Z\\\",\\\"message\\\":\\\"0:16:25.017976 6145 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 10:16:25.018126 6145 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 10:16:25.018310 6145 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:25.018453 6145 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018512 6145 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018817 6145 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 10:16:25.018867 6145 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 10:16:25.018886 6145 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 10:16:25.018891 6145 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 10:16:25.018920 6145 factory.go:656] Stopping watch factory\\\\nI1204 10:16:25.018935 6145 ovnkube.go:599] Stopped ovnkube\\\\nI1204 10:16:25.018949 6145 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 10:16:25.018966 6145 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 10:16:25.018961 6145 handler.go:208] Removed *v1.Pod event ha\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.322630 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:28 crc kubenswrapper[4943]: E1204 10:16:28.322695 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.333187 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.345158 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.357775 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.359663 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.359695 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.359705 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.359720 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.359730 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:28Z","lastTransitionTime":"2025-12-04T10:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.372223 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.374831 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.374869 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpblt\" (UniqueName: \"kubernetes.io/projected/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-kube-api-access-bpblt\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.389177 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.400083 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.413172 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.425160 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.439562 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.451940 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.462163 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.462226 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.462238 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.462256 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.462266 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:28Z","lastTransitionTime":"2025-12-04T10:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.466651 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.475361 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpblt\" (UniqueName: \"kubernetes.io/projected/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-kube-api-access-bpblt\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.475436 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:28 crc kubenswrapper[4943]: E1204 10:16:28.475538 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:16:28 crc kubenswrapper[4943]: E1204 10:16:28.475587 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs podName:eb85a8e0-6c90-47cb-a1b5-9ecd5244c710 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:28.975573966 +0000 UTC m=+37.564549834 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs") pod "network-metrics-daemon-hw5pj" (UID: "eb85a8e0-6c90-47cb-a1b5-9ecd5244c710") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.479065 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.489141 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.494443 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpblt\" (UniqueName: \"kubernetes.io/projected/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-kube-api-access-bpblt\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.501499 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.515939 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.530600 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.541387 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.552502 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.565076 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.565131 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.565145 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.565169 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.565186 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:28Z","lastTransitionTime":"2025-12-04T10:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.565316 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.584044 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:25Z\\\",\\\"message\\\":\\\"0:16:25.017976 6145 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 10:16:25.018126 6145 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 10:16:25.018310 6145 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:25.018453 6145 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018512 6145 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018817 6145 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 10:16:25.018867 6145 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 10:16:25.018886 6145 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 10:16:25.018891 6145 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 10:16:25.018920 6145 factory.go:656] Stopping watch factory\\\\nI1204 10:16:25.018935 6145 ovnkube.go:599] Stopped ovnkube\\\\nI1204 10:16:25.018949 6145 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 10:16:25.018966 6145 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 10:16:25.018961 6145 handler.go:208] Removed *v1.Pod event ha\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"message\\\":\\\" or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z]\\\\nI1204 10:16:26.015295 6312 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015300 6312 services_controller.go:453] Built service openshift-operator-lifecycle-manager/catalog-operator-metrics template LB for network=default: []services.LB{}\\\\nI1204 10:16:26.015303 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015304 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-7g4wf\\\\nI1204 10:16:26.015313 6312 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 10:16:26.015235 6312 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0079b22bd 0xc0079b22be}] [] []},Spec:ServiceSpec{Ports:[]ServicePo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.594415 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.607907 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.625207 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.640482 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.652764 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.667654 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.667674 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.667720 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.667734 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.667754 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.667764 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:28Z","lastTransitionTime":"2025-12-04T10:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.681011 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:28Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.770383 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.770453 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.770471 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.770498 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.770518 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:28Z","lastTransitionTime":"2025-12-04T10:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.872260 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.872323 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.872335 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.872351 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.872361 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:28Z","lastTransitionTime":"2025-12-04T10:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.973917 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.973964 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.973982 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.973998 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.974008 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:28Z","lastTransitionTime":"2025-12-04T10:16:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:28 crc kubenswrapper[4943]: I1204 10:16:28.979475 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:28 crc kubenswrapper[4943]: E1204 10:16:28.979676 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:16:28 crc kubenswrapper[4943]: E1204 10:16:28.979774 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs podName:eb85a8e0-6c90-47cb-a1b5-9ecd5244c710 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:29.979751961 +0000 UTC m=+38.568727869 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs") pod "network-metrics-daemon-hw5pj" (UID: "eb85a8e0-6c90-47cb-a1b5-9ecd5244c710") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.076094 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.076454 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.076466 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.076487 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.076498 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:29Z","lastTransitionTime":"2025-12-04T10:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.179360 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.179406 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.179419 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.179436 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.179448 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:29Z","lastTransitionTime":"2025-12-04T10:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.181794 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.181970 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.181984 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:16:45.181960188 +0000 UTC m=+53.770936056 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.182060 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.182097 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:45.182089951 +0000 UTC m=+53.771065819 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.182176 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.182214 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.182227 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.182273 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:45.182259926 +0000 UTC m=+53.771235794 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.182059 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.182325 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.182402 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.182431 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:45.182424729 +0000 UTC m=+53.771400597 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.260114 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/1.log" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.281781 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.281822 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.281834 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.281849 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.281859 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:29Z","lastTransitionTime":"2025-12-04T10:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.383658 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.383818 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.383838 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.383851 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.384169 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:45.384153495 +0000 UTC m=+53.973129363 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.384253 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.384286 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.384296 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.384311 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.384322 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:29Z","lastTransitionTime":"2025-12-04T10:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.487026 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.487091 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.487133 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.487167 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.487191 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:29Z","lastTransitionTime":"2025-12-04T10:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.564113 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.564109 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.564316 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.564341 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.564263 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.564408 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.564129 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.564454 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.589679 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.589719 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.589732 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.589750 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.589761 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:29Z","lastTransitionTime":"2025-12-04T10:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.692550 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.692593 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.692601 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.692615 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.692624 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:29Z","lastTransitionTime":"2025-12-04T10:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.794961 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.795000 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.795009 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.795023 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.795032 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:29Z","lastTransitionTime":"2025-12-04T10:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.898925 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.898974 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.898988 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.899005 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.899016 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:29Z","lastTransitionTime":"2025-12-04T10:16:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:29 crc kubenswrapper[4943]: I1204 10:16:29.988146 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.988388 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:16:29 crc kubenswrapper[4943]: E1204 10:16:29.988492 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs podName:eb85a8e0-6c90-47cb-a1b5-9ecd5244c710 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:31.988468921 +0000 UTC m=+40.577444799 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs") pod "network-metrics-daemon-hw5pj" (UID: "eb85a8e0-6c90-47cb-a1b5-9ecd5244c710") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.001126 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.001163 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.001173 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.001190 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.001204 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:30Z","lastTransitionTime":"2025-12-04T10:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.104356 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.104433 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.104446 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.104463 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.104473 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:30Z","lastTransitionTime":"2025-12-04T10:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.207334 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.207385 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.207394 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.207408 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.207416 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:30Z","lastTransitionTime":"2025-12-04T10:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.310363 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.310408 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.310435 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.310451 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.310461 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:30Z","lastTransitionTime":"2025-12-04T10:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.413361 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.413402 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.413412 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.413427 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.413436 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:30Z","lastTransitionTime":"2025-12-04T10:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.515982 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.516051 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.516066 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.516083 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.516094 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:30Z","lastTransitionTime":"2025-12-04T10:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.617751 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.617803 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.617815 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.617832 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.617845 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:30Z","lastTransitionTime":"2025-12-04T10:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.720666 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.720730 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.720741 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.720756 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.720765 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:30Z","lastTransitionTime":"2025-12-04T10:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.823071 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.823129 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.823140 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.823160 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.823173 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:30Z","lastTransitionTime":"2025-12-04T10:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.926022 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.926065 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.926079 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.926096 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:30 crc kubenswrapper[4943]: I1204 10:16:30.926105 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:30Z","lastTransitionTime":"2025-12-04T10:16:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.028787 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.028828 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.028839 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.028852 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.028860 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:31Z","lastTransitionTime":"2025-12-04T10:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.131753 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.131825 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.131899 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.131932 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.131954 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:31Z","lastTransitionTime":"2025-12-04T10:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.233941 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.233983 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.233998 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.234013 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.234026 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:31Z","lastTransitionTime":"2025-12-04T10:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.336539 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.336583 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.336594 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.336607 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.336617 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:31Z","lastTransitionTime":"2025-12-04T10:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.444966 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.445763 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.445799 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.445821 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.445836 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:31Z","lastTransitionTime":"2025-12-04T10:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.548245 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.548332 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.548348 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.548374 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.548390 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:31Z","lastTransitionTime":"2025-12-04T10:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.564693 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:31 crc kubenswrapper[4943]: E1204 10:16:31.564923 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.565333 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.565374 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.565448 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:31 crc kubenswrapper[4943]: E1204 10:16:31.565448 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:31 crc kubenswrapper[4943]: E1204 10:16:31.565572 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:31 crc kubenswrapper[4943]: E1204 10:16:31.565689 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.650931 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.651004 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.651018 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.651036 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.651048 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:31Z","lastTransitionTime":"2025-12-04T10:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.754515 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.754565 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.754576 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.754592 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.754602 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:31Z","lastTransitionTime":"2025-12-04T10:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.857364 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.857404 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.857413 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.857430 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.857439 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:31Z","lastTransitionTime":"2025-12-04T10:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.960270 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.960327 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.960340 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.960360 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:31 crc kubenswrapper[4943]: I1204 10:16:31.960373 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:31Z","lastTransitionTime":"2025-12-04T10:16:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.010855 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:32 crc kubenswrapper[4943]: E1204 10:16:32.011068 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:16:32 crc kubenswrapper[4943]: E1204 10:16:32.011187 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs podName:eb85a8e0-6c90-47cb-a1b5-9ecd5244c710 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:36.011152737 +0000 UTC m=+44.600128645 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs") pod "network-metrics-daemon-hw5pj" (UID: "eb85a8e0-6c90-47cb-a1b5-9ecd5244c710") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.063982 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.064050 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.064061 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.064082 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.064094 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:32Z","lastTransitionTime":"2025-12-04T10:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.167256 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.167311 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.167321 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.167339 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.167348 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:32Z","lastTransitionTime":"2025-12-04T10:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.270073 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.270124 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.270135 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.270152 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.270163 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:32Z","lastTransitionTime":"2025-12-04T10:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.372835 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.372906 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.372921 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.372945 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.372958 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:32Z","lastTransitionTime":"2025-12-04T10:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.475559 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.475635 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.475647 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.475663 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.475674 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:32Z","lastTransitionTime":"2025-12-04T10:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.577763 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.577807 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.577826 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.577845 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.577857 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:32Z","lastTransitionTime":"2025-12-04T10:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.579610 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.593469 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.607076 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.635934 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:25Z\\\",\\\"message\\\":\\\"0:16:25.017976 6145 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 10:16:25.018126 6145 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 10:16:25.018310 6145 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:25.018453 6145 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018512 6145 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018817 6145 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 10:16:25.018867 6145 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 10:16:25.018886 6145 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 10:16:25.018891 6145 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 10:16:25.018920 6145 factory.go:656] Stopping watch factory\\\\nI1204 10:16:25.018935 6145 ovnkube.go:599] Stopped ovnkube\\\\nI1204 10:16:25.018949 6145 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 10:16:25.018966 6145 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 10:16:25.018961 6145 handler.go:208] Removed *v1.Pod event ha\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"message\\\":\\\" or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z]\\\\nI1204 10:16:26.015295 6312 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015300 6312 services_controller.go:453] Built service openshift-operator-lifecycle-manager/catalog-operator-metrics template LB for network=default: []services.LB{}\\\\nI1204 10:16:26.015303 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015304 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-7g4wf\\\\nI1204 10:16:26.015313 6312 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 10:16:26.015235 6312 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0079b22bd 0xc0079b22be}] [] []},Spec:ServiceSpec{Ports:[]ServicePo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.651656 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.665863 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.676064 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.679759 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.679794 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.679805 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.679822 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.679836 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:32Z","lastTransitionTime":"2025-12-04T10:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.687245 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.700927 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.713074 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.723622 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.734500 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.745296 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.762175 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.776228 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.781725 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.781789 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.781801 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.781821 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.781830 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:32Z","lastTransitionTime":"2025-12-04T10:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.793163 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:32Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.885232 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.885277 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.885314 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.885331 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.885342 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:32Z","lastTransitionTime":"2025-12-04T10:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.987809 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.987874 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.987882 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.987897 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:32 crc kubenswrapper[4943]: I1204 10:16:32.987907 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:32Z","lastTransitionTime":"2025-12-04T10:16:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.090641 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.090676 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.090684 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.090697 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.090708 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:33Z","lastTransitionTime":"2025-12-04T10:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.193104 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.193153 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.193170 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.193192 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.193209 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:33Z","lastTransitionTime":"2025-12-04T10:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.295932 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.295972 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.295983 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.295996 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.296006 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:33Z","lastTransitionTime":"2025-12-04T10:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.398184 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.398243 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.398254 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.398270 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.398280 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:33Z","lastTransitionTime":"2025-12-04T10:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.500642 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.500754 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.500778 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.500811 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.500831 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:33Z","lastTransitionTime":"2025-12-04T10:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.563637 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.563683 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.563711 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.563668 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:33 crc kubenswrapper[4943]: E1204 10:16:33.563795 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:33 crc kubenswrapper[4943]: E1204 10:16:33.563862 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:33 crc kubenswrapper[4943]: E1204 10:16:33.563921 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:33 crc kubenswrapper[4943]: E1204 10:16:33.563968 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.603340 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.603380 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.603390 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.603408 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.603419 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:33Z","lastTransitionTime":"2025-12-04T10:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.705871 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.705919 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.705928 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.705943 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.705954 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:33Z","lastTransitionTime":"2025-12-04T10:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.808529 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.808569 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.808578 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.808595 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.808607 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:33Z","lastTransitionTime":"2025-12-04T10:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.911270 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.911317 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.911327 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.911342 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:33 crc kubenswrapper[4943]: I1204 10:16:33.911351 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:33Z","lastTransitionTime":"2025-12-04T10:16:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.013746 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.013791 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.013800 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.013818 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.013827 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.041985 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.042038 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.042047 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.042061 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.042072 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: E1204 10:16:34.054259 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:34Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.059261 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.059309 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.059322 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.059339 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.059350 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: E1204 10:16:34.070822 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:34Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.074921 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.074974 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.074993 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.075014 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.075024 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: E1204 10:16:34.088670 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:34Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.093162 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.093224 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.093237 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.093257 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.093266 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: E1204 10:16:34.107010 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:34Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.114842 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.114883 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.114892 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.114906 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.114916 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: E1204 10:16:34.127416 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:34Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:34 crc kubenswrapper[4943]: E1204 10:16:34.127553 4943 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.128938 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.128973 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.128987 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.129001 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.129012 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.232660 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.232729 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.232739 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.232754 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.232763 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.335137 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.335236 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.335255 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.335275 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.335289 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.438244 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.438311 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.438327 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.438349 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.438366 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.542320 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.542355 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.542365 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.542379 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.542387 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.644262 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.644313 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.644326 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.644343 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.644356 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.747044 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.747113 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.747134 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.747158 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.747175 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.849412 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.849455 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.849463 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.849476 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.849484 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.951373 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.951417 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.951428 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.951443 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:34 crc kubenswrapper[4943]: I1204 10:16:34.951464 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:34Z","lastTransitionTime":"2025-12-04T10:16:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.054153 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.054191 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.054217 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.054233 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.054245 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:35Z","lastTransitionTime":"2025-12-04T10:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.155699 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.155766 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.155778 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.155793 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.155803 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:35Z","lastTransitionTime":"2025-12-04T10:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.258318 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.258390 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.258412 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.258440 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.258463 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:35Z","lastTransitionTime":"2025-12-04T10:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.360942 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.360993 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.361013 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.361032 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.361044 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:35Z","lastTransitionTime":"2025-12-04T10:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.463876 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.463925 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.463942 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.463961 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.463972 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:35Z","lastTransitionTime":"2025-12-04T10:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.563868 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.563868 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:35 crc kubenswrapper[4943]: E1204 10:16:35.564011 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.563980 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:35 crc kubenswrapper[4943]: E1204 10:16:35.564071 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.563902 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:35 crc kubenswrapper[4943]: E1204 10:16:35.564112 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:35 crc kubenswrapper[4943]: E1204 10:16:35.564307 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.565669 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.565722 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.565734 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.565749 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.565761 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:35Z","lastTransitionTime":"2025-12-04T10:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.668100 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.668160 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.668171 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.668190 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.668219 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:35Z","lastTransitionTime":"2025-12-04T10:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.770409 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.770476 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.770492 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.770519 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.770536 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:35Z","lastTransitionTime":"2025-12-04T10:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.873631 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.873699 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.873712 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.873731 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.873745 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:35Z","lastTransitionTime":"2025-12-04T10:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.976716 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.976775 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.976792 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.976814 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:35 crc kubenswrapper[4943]: I1204 10:16:35.976833 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:35Z","lastTransitionTime":"2025-12-04T10:16:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.058232 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:36 crc kubenswrapper[4943]: E1204 10:16:36.058552 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:16:36 crc kubenswrapper[4943]: E1204 10:16:36.058668 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs podName:eb85a8e0-6c90-47cb-a1b5-9ecd5244c710 nodeName:}" failed. No retries permitted until 2025-12-04 10:16:44.058646431 +0000 UTC m=+52.647622299 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs") pod "network-metrics-daemon-hw5pj" (UID: "eb85a8e0-6c90-47cb-a1b5-9ecd5244c710") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.079330 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.079381 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.079392 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.079409 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.079420 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:36Z","lastTransitionTime":"2025-12-04T10:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.182679 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.182736 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.182748 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.182767 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.182778 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:36Z","lastTransitionTime":"2025-12-04T10:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.286188 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.286308 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.286321 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.286343 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.286358 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:36Z","lastTransitionTime":"2025-12-04T10:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.388924 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.389004 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.389018 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.389043 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.389059 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:36Z","lastTransitionTime":"2025-12-04T10:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.492410 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.492471 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.492488 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.492515 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.492529 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:36Z","lastTransitionTime":"2025-12-04T10:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.595390 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.595460 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.595476 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.595502 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.595519 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:36Z","lastTransitionTime":"2025-12-04T10:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.697921 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.697994 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.698009 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.698026 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.698036 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:36Z","lastTransitionTime":"2025-12-04T10:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.800984 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.801049 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.801066 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.801084 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.801096 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:36Z","lastTransitionTime":"2025-12-04T10:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.903714 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.903783 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.903806 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.903833 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:36 crc kubenswrapper[4943]: I1204 10:16:36.903854 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:36Z","lastTransitionTime":"2025-12-04T10:16:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.006715 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.006760 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.006780 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.006802 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.006814 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:37Z","lastTransitionTime":"2025-12-04T10:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.110453 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.110510 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.110522 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.110542 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.110557 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:37Z","lastTransitionTime":"2025-12-04T10:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.213009 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.213064 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.213079 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.213096 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.213109 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:37Z","lastTransitionTime":"2025-12-04T10:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.316504 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.316560 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.316580 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.316600 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.316615 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:37Z","lastTransitionTime":"2025-12-04T10:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.419330 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.419370 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.419383 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.419397 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.419407 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:37Z","lastTransitionTime":"2025-12-04T10:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.522036 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.522097 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.522113 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.522136 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.522156 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:37Z","lastTransitionTime":"2025-12-04T10:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.563749 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.563829 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.563811 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.563749 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:37 crc kubenswrapper[4943]: E1204 10:16:37.563972 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:37 crc kubenswrapper[4943]: E1204 10:16:37.564001 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:37 crc kubenswrapper[4943]: E1204 10:16:37.564066 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:37 crc kubenswrapper[4943]: E1204 10:16:37.564127 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.625508 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.625563 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.625576 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.625598 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.625623 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:37Z","lastTransitionTime":"2025-12-04T10:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.728924 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.728978 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.728991 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.729008 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.729020 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:37Z","lastTransitionTime":"2025-12-04T10:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.831812 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.831880 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.831896 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.831921 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.831938 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:37Z","lastTransitionTime":"2025-12-04T10:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.934614 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.934671 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.934681 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.934696 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:37 crc kubenswrapper[4943]: I1204 10:16:37.934705 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:37Z","lastTransitionTime":"2025-12-04T10:16:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.037655 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.037710 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.037728 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.037749 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.037764 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:38Z","lastTransitionTime":"2025-12-04T10:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.140646 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.140711 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.140724 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.140746 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.140757 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:38Z","lastTransitionTime":"2025-12-04T10:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.242905 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.242958 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.242969 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.242986 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.242998 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:38Z","lastTransitionTime":"2025-12-04T10:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.345491 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.345549 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.345566 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.345588 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.345606 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:38Z","lastTransitionTime":"2025-12-04T10:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.448890 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.448961 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.448974 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.448989 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.448998 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:38Z","lastTransitionTime":"2025-12-04T10:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.551808 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.551871 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.551887 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.551904 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.551914 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:38Z","lastTransitionTime":"2025-12-04T10:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.654472 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.654537 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.654554 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.654582 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.654599 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:38Z","lastTransitionTime":"2025-12-04T10:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.757022 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.757073 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.757086 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.757105 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.757123 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:38Z","lastTransitionTime":"2025-12-04T10:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.859023 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.859065 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.859074 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.859087 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.859098 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:38Z","lastTransitionTime":"2025-12-04T10:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.961370 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.961411 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.961422 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.961437 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:38 crc kubenswrapper[4943]: I1204 10:16:38.961449 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:38Z","lastTransitionTime":"2025-12-04T10:16:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.064767 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.064878 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.064897 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.064923 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.064949 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:39Z","lastTransitionTime":"2025-12-04T10:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.168602 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.168640 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.168660 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.168674 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.168683 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:39Z","lastTransitionTime":"2025-12-04T10:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.271248 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.271296 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.271306 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.271322 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.271332 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:39Z","lastTransitionTime":"2025-12-04T10:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.373689 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.373728 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.373744 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.373759 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.373770 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:39Z","lastTransitionTime":"2025-12-04T10:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.476357 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.476401 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.476419 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.476436 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.476447 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:39Z","lastTransitionTime":"2025-12-04T10:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.564074 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.564124 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.564123 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.564089 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:39 crc kubenswrapper[4943]: E1204 10:16:39.564292 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:39 crc kubenswrapper[4943]: E1204 10:16:39.564318 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:39 crc kubenswrapper[4943]: E1204 10:16:39.564372 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:39 crc kubenswrapper[4943]: E1204 10:16:39.564438 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.578924 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.578971 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.578985 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.579000 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.579010 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:39Z","lastTransitionTime":"2025-12-04T10:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.681191 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.681253 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.681265 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.681313 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.681325 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:39Z","lastTransitionTime":"2025-12-04T10:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.783637 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.783711 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.783736 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.783766 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.783789 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:39Z","lastTransitionTime":"2025-12-04T10:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.888225 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.888274 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.888292 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.888310 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.888319 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:39Z","lastTransitionTime":"2025-12-04T10:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.991062 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.991114 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.991127 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.991146 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:39 crc kubenswrapper[4943]: I1204 10:16:39.991160 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:39Z","lastTransitionTime":"2025-12-04T10:16:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.094140 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.094192 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.094219 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.094235 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.094246 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:40Z","lastTransitionTime":"2025-12-04T10:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.196374 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.196430 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.196442 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.196461 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.196477 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:40Z","lastTransitionTime":"2025-12-04T10:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.298745 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.298799 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.298815 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.298832 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.298845 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:40Z","lastTransitionTime":"2025-12-04T10:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.401573 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.401625 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.401638 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.401654 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.401667 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:40Z","lastTransitionTime":"2025-12-04T10:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.503851 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.503897 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.503913 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.503930 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.503939 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:40Z","lastTransitionTime":"2025-12-04T10:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.606702 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.606757 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.606766 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.606780 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.606790 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:40Z","lastTransitionTime":"2025-12-04T10:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.709486 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.709586 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.709607 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.709624 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.709635 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:40Z","lastTransitionTime":"2025-12-04T10:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.812515 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.812588 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.812607 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.812638 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.812658 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:40Z","lastTransitionTime":"2025-12-04T10:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.915375 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.915434 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.915446 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.915461 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:40 crc kubenswrapper[4943]: I1204 10:16:40.915472 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:40Z","lastTransitionTime":"2025-12-04T10:16:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.017861 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.017925 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.017936 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.017958 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.017970 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:41Z","lastTransitionTime":"2025-12-04T10:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.120959 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.121005 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.121014 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.121027 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.121038 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:41Z","lastTransitionTime":"2025-12-04T10:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.224067 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.224110 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.224120 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.224149 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.224158 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:41Z","lastTransitionTime":"2025-12-04T10:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.327944 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.328022 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.328035 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.328055 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.328073 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:41Z","lastTransitionTime":"2025-12-04T10:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.430362 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.430425 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.430436 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.430453 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.430464 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:41Z","lastTransitionTime":"2025-12-04T10:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.533385 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.533450 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.533458 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.533473 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.533484 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:41Z","lastTransitionTime":"2025-12-04T10:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.563696 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.563729 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.563782 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.563793 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:41 crc kubenswrapper[4943]: E1204 10:16:41.563915 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:41 crc kubenswrapper[4943]: E1204 10:16:41.564050 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:41 crc kubenswrapper[4943]: E1204 10:16:41.564189 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:41 crc kubenswrapper[4943]: E1204 10:16:41.564395 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.635907 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.635966 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.635980 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.636000 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.636014 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:41Z","lastTransitionTime":"2025-12-04T10:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.738554 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.738606 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.738617 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.738633 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.738644 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:41Z","lastTransitionTime":"2025-12-04T10:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.841901 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.841952 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.841962 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.841979 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.841992 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:41Z","lastTransitionTime":"2025-12-04T10:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.944158 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.944231 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.944244 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.944261 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:41 crc kubenswrapper[4943]: I1204 10:16:41.944274 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:41Z","lastTransitionTime":"2025-12-04T10:16:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.046900 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.046943 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.046952 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.046968 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.046978 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:42Z","lastTransitionTime":"2025-12-04T10:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.149573 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.149646 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.149670 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.149701 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.149723 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:42Z","lastTransitionTime":"2025-12-04T10:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.251648 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.251691 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.251700 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.251714 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.251724 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:42Z","lastTransitionTime":"2025-12-04T10:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.353721 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.353773 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.353787 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.353802 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.353812 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:42Z","lastTransitionTime":"2025-12-04T10:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.455878 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.455934 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.455942 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.455958 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.455967 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:42Z","lastTransitionTime":"2025-12-04T10:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.557876 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.557936 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.557959 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.557981 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.557998 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:42Z","lastTransitionTime":"2025-12-04T10:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.582035 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.598631 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.613222 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.624285 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.642069 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a17adc2a8012f37de675ea90e47e1f0c4bd2ea2d9d777417caea6ca553d3c89\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:25Z\\\",\\\"message\\\":\\\"0:16:25.017976 6145 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 10:16:25.018126 6145 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 10:16:25.018310 6145 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:25.018453 6145 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018512 6145 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:25.018817 6145 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 10:16:25.018867 6145 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 10:16:25.018886 6145 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1204 10:16:25.018891 6145 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1204 10:16:25.018920 6145 factory.go:656] Stopping watch factory\\\\nI1204 10:16:25.018935 6145 ovnkube.go:599] Stopped ovnkube\\\\nI1204 10:16:25.018949 6145 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 10:16:25.018966 6145 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 10:16:25.018961 6145 handler.go:208] Removed *v1.Pod event ha\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"message\\\":\\\" or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z]\\\\nI1204 10:16:26.015295 6312 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015300 6312 services_controller.go:453] Built service openshift-operator-lifecycle-manager/catalog-operator-metrics template LB for network=default: []services.LB{}\\\\nI1204 10:16:26.015303 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015304 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-7g4wf\\\\nI1204 10:16:26.015313 6312 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 10:16:26.015235 6312 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0079b22bd 0xc0079b22be}] [] []},Spec:ServiceSpec{Ports:[]ServicePo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.658924 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.659957 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.659987 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.659996 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.660010 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.660020 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:42Z","lastTransitionTime":"2025-12-04T10:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.670859 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.682400 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.693512 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.703798 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.712907 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.722697 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.733619 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.746516 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.758732 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.762352 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.762391 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.762399 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.762413 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.762422 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:42Z","lastTransitionTime":"2025-12-04T10:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.769967 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:42Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.865248 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.865300 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.865317 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.865340 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.865356 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:42Z","lastTransitionTime":"2025-12-04T10:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.967344 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.967387 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.967396 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.967413 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:42 crc kubenswrapper[4943]: I1204 10:16:42.967422 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:42Z","lastTransitionTime":"2025-12-04T10:16:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.069471 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.069545 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.069572 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.069604 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.069625 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:43Z","lastTransitionTime":"2025-12-04T10:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.172549 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.172649 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.172672 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.172702 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.172728 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:43Z","lastTransitionTime":"2025-12-04T10:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.275949 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.276012 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.276022 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.276047 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.276060 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:43Z","lastTransitionTime":"2025-12-04T10:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.379482 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.379544 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.379567 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.379590 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.379608 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:43Z","lastTransitionTime":"2025-12-04T10:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.481929 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.481985 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.481996 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.482012 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.482023 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:43Z","lastTransitionTime":"2025-12-04T10:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.563728 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.563797 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.563753 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:43 crc kubenswrapper[4943]: E1204 10:16:43.563892 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.563892 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:43 crc kubenswrapper[4943]: E1204 10:16:43.564261 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:43 crc kubenswrapper[4943]: E1204 10:16:43.564425 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.564477 4943 scope.go:117] "RemoveContainer" containerID="da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1" Dec 04 10:16:43 crc kubenswrapper[4943]: E1204 10:16:43.564526 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.580194 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.585328 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.585875 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.585965 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.586061 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.586147 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:43Z","lastTransitionTime":"2025-12-04T10:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.596839 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.611153 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.623575 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.639746 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.649417 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.660019 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.670537 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.681681 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.688840 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.688975 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.689057 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.689142 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.689284 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:43Z","lastTransitionTime":"2025-12-04T10:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.694263 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.706632 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.722951 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.733936 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.745183 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.758709 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.780217 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"message\\\":\\\" or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z]\\\\nI1204 10:16:26.015295 6312 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015300 6312 services_controller.go:453] Built service openshift-operator-lifecycle-manager/catalog-operator-metrics template LB for network=default: []services.LB{}\\\\nI1204 10:16:26.015303 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015304 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-7g4wf\\\\nI1204 10:16:26.015313 6312 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 10:16:26.015235 6312 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0079b22bd 0xc0079b22be}] [] []},Spec:ServiceSpec{Ports:[]ServicePo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:43Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.791696 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.791745 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.791756 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.791771 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.791780 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:43Z","lastTransitionTime":"2025-12-04T10:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.894092 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.894134 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.894147 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.894162 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.894172 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:43Z","lastTransitionTime":"2025-12-04T10:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.996565 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.996612 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.996623 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.996637 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:43 crc kubenswrapper[4943]: I1204 10:16:43.996647 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:43Z","lastTransitionTime":"2025-12-04T10:16:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.008885 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.018290 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.021051 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.032598 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.052918 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.067068 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.079288 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.088278 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.097550 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.098388 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.098420 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.098432 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.098448 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.098459 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.110629 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.129898 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"message\\\":\\\" or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z]\\\\nI1204 10:16:26.015295 6312 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015300 6312 services_controller.go:453] Built service openshift-operator-lifecycle-manager/catalog-operator-metrics template LB for network=default: []services.LB{}\\\\nI1204 10:16:26.015303 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015304 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-7g4wf\\\\nI1204 10:16:26.015313 6312 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 10:16:26.015235 6312 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0079b22bd 0xc0079b22be}] [] []},Spec:ServiceSpec{Ports:[]ServicePo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.141112 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.142478 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:44 crc kubenswrapper[4943]: E1204 10:16:44.142653 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:16:44 crc kubenswrapper[4943]: E1204 10:16:44.142731 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs podName:eb85a8e0-6c90-47cb-a1b5-9ecd5244c710 nodeName:}" failed. No retries permitted until 2025-12-04 10:17:00.142713165 +0000 UTC m=+68.731689033 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs") pod "network-metrics-daemon-hw5pj" (UID: "eb85a8e0-6c90-47cb-a1b5-9ecd5244c710") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.152508 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.162907 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.174485 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.187562 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.200269 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.200520 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.200616 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.200755 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.200845 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.203504 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.216050 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.303663 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.303935 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.304128 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.304234 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.304314 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.318744 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/1.log" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.321973 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.327306 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.327350 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.327360 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.327377 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.327387 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.338065 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: E1204 10:16:44.338454 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.341436 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.341475 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.341488 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.341505 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.341519 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.350334 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: E1204 10:16:44.352387 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.355146 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.355175 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.355185 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.355225 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.355240 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.362677 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: E1204 10:16:44.369007 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.372823 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.372872 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.372883 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.372899 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.372914 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.380082 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: E1204 10:16:44.384424 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.387897 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.387940 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.387953 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.387971 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.387983 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.393271 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: E1204 10:16:44.400875 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: E1204 10:16:44.401120 4943 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.403759 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.407656 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.407767 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.407858 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.407934 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.408018 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.415912 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.428336 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.442464 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.459111 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.511045 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.511096 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.511111 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.511132 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.511152 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.614384 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.614448 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.614464 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.614484 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.614497 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.619099 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.637844 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.657562 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b54b12fc-f44c-42cf-a647-6ff44c005e5b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07144e00536ba9c9e2f8f9be3199ca05f6b6a864d5ee1fe01b49c4e30a244ae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e348e3b6d23732341818a2d444fc18b715cc7729d49508ba0bec8a349118aba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61720056d847f2e14a4c17ca79e24d8e330530b0ed68c256aa2ddcb35552c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.673255 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.690592 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"message\\\":\\\" or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z]\\\\nI1204 10:16:26.015295 6312 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015300 6312 services_controller.go:453] Built service openshift-operator-lifecycle-manager/catalog-operator-metrics template LB for network=default: []services.LB{}\\\\nI1204 10:16:26.015303 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015304 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-7g4wf\\\\nI1204 10:16:26.015313 6312 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 10:16:26.015235 6312 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0079b22bd 0xc0079b22be}] [] []},Spec:ServiceSpec{Ports:[]ServicePo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.702442 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.713424 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:44Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.717713 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.717769 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.717781 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.717804 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.717817 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.820307 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.820362 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.820374 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.820391 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.820402 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.922937 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.923007 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.923023 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.923048 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:44 crc kubenswrapper[4943]: I1204 10:16:44.923067 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:44Z","lastTransitionTime":"2025-12-04T10:16:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.025553 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.025602 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.025614 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.025635 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.025647 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:45Z","lastTransitionTime":"2025-12-04T10:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.127970 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.128020 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.128031 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.128048 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.128061 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:45Z","lastTransitionTime":"2025-12-04T10:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.230175 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.230254 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.230267 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.230283 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.230315 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:45Z","lastTransitionTime":"2025-12-04T10:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.255876 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.255998 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.256043 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:17:17.256020222 +0000 UTC m=+85.844996090 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.256106 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.256120 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.256137 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.256139 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.256148 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.256190 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.256239 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:17:17.256230697 +0000 UTC m=+85.845206565 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.256251 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 10:17:17.256246368 +0000 UTC m=+85.845222236 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.256311 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.256423 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:17:17.256396641 +0000 UTC m=+85.845372609 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.332083 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.332133 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.332144 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.332160 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.332171 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:45Z","lastTransitionTime":"2025-12-04T10:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.434961 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.435013 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.435025 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.435041 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.435052 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:45Z","lastTransitionTime":"2025-12-04T10:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.458606 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.458771 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.458795 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.458806 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.458856 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 10:17:17.458843404 +0000 UTC m=+86.047819272 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.537686 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.537732 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.537742 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.537760 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.537773 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:45Z","lastTransitionTime":"2025-12-04T10:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.564594 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.564626 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.564646 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.564594 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.564729 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.564789 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.564866 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:45 crc kubenswrapper[4943]: E1204 10:16:45.564954 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.640716 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.640765 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.640774 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.640791 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.640801 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:45Z","lastTransitionTime":"2025-12-04T10:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.744290 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.744353 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.744370 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.744393 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.744411 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:45Z","lastTransitionTime":"2025-12-04T10:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.848055 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.848136 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.848160 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.848238 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.848274 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:45Z","lastTransitionTime":"2025-12-04T10:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.950952 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.950992 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.951000 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.951017 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:45 crc kubenswrapper[4943]: I1204 10:16:45.951026 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:45Z","lastTransitionTime":"2025-12-04T10:16:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.054138 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.054185 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.054195 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.054238 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.054251 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:46Z","lastTransitionTime":"2025-12-04T10:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.156986 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.157044 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.157082 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.157101 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.157114 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:46Z","lastTransitionTime":"2025-12-04T10:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.259330 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.259375 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.259383 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.259397 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.259406 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:46Z","lastTransitionTime":"2025-12-04T10:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.363257 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.363305 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.363317 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.363336 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.363346 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:46Z","lastTransitionTime":"2025-12-04T10:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.469842 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.469965 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.469978 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.469994 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.470015 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:46Z","lastTransitionTime":"2025-12-04T10:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.572359 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.572388 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.572396 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.572407 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.572416 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:46Z","lastTransitionTime":"2025-12-04T10:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.674876 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.674923 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.674936 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.674955 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.674968 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:46Z","lastTransitionTime":"2025-12-04T10:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.778229 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.778611 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.778631 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.778654 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.778667 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:46Z","lastTransitionTime":"2025-12-04T10:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.881269 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.881317 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.881328 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.881358 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.881369 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:46Z","lastTransitionTime":"2025-12-04T10:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.984481 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.984532 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.984543 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.984560 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:46 crc kubenswrapper[4943]: I1204 10:16:46.984571 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:46Z","lastTransitionTime":"2025-12-04T10:16:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.087124 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.087174 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.087190 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.087227 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.087243 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:47Z","lastTransitionTime":"2025-12-04T10:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.189870 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.189928 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.189942 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.189963 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.189974 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:47Z","lastTransitionTime":"2025-12-04T10:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.292512 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.292549 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.292557 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.292570 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.292578 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:47Z","lastTransitionTime":"2025-12-04T10:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.334235 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/2.log" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.335225 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/1.log" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.338504 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36" exitCode=1 Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.338576 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36"} Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.338672 4943 scope.go:117] "RemoveContainer" containerID="da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.341225 4943 scope.go:117] "RemoveContainer" containerID="7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36" Dec 04 10:16:47 crc kubenswrapper[4943]: E1204 10:16:47.345742 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.366121 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.382180 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.395780 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.395823 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.395844 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.395912 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.395933 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:47Z","lastTransitionTime":"2025-12-04T10:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.395790 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.417435 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.427152 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.437610 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.452917 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.464679 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.476412 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.490941 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.499914 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.499958 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.499972 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.499993 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.500012 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:47Z","lastTransitionTime":"2025-12-04T10:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.506897 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b54b12fc-f44c-42cf-a647-6ff44c005e5b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07144e00536ba9c9e2f8f9be3199ca05f6b6a864d5ee1fe01b49c4e30a244ae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e348e3b6d23732341818a2d444fc18b715cc7729d49508ba0bec8a349118aba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61720056d847f2e14a4c17ca79e24d8e330530b0ed68c256aa2ddcb35552c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.520361 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.535187 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.545988 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.556498 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.564179 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.564179 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:47 crc kubenswrapper[4943]: E1204 10:16:47.564351 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.564223 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:47 crc kubenswrapper[4943]: E1204 10:16:47.564412 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.564179 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:47 crc kubenswrapper[4943]: E1204 10:16:47.564469 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:47 crc kubenswrapper[4943]: E1204 10:16:47.564508 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.571334 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.593717 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://da30555d2df55b723bd660acd9de40e2d3eeb4d0bd2360b9973f4d6c96d33fb1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"message\\\":\\\" or is not yet valid: current time 2025-12-04T10:16:25Z is after 2025-08-24T17:21:41Z]\\\\nI1204 10:16:26.015295 6312 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015300 6312 services_controller.go:453] Built service openshift-operator-lifecycle-manager/catalog-operator-metrics template LB for network=default: []services.LB{}\\\\nI1204 10:16:26.015303 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1204 10:16:26.015304 6312 obj_retry.go:365] Adding new object: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-7g4wf\\\\nI1204 10:16:26.015313 6312 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1204 10:16:26.015235 6312 services_controller.go:434] Service openshift-operator-lifecycle-manager/packageserver-service retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{packageserver-service openshift-operator-lifecycle-manager a60a1f74-c6ff-4c81-96ae-27ba9796ba61 5485 0 2025-02-23 05:23:24 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[olm.managed:true] map[] [{operators.coreos.com/v1alpha1 ClusterServiceVersion packageserver bbc08db6-5ba4-4fc4-b49d-26331e1e728b 0xc0079b22bd 0xc0079b22be}] [] []},Spec:ServiceSpec{Ports:[]ServicePo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:46Z\\\",\\\"message\\\":\\\"opping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:45.159276 6525 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159565 6525 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159786 6525 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159885 6525 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159901 6525 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159949 6525 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159979 6525 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.160402 6525 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:47Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.602861 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.602938 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.602950 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.602968 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.602980 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:47Z","lastTransitionTime":"2025-12-04T10:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.705634 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.705673 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.705684 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.705704 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.705713 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:47Z","lastTransitionTime":"2025-12-04T10:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.808507 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.808547 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.808556 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.808573 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.808584 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:47Z","lastTransitionTime":"2025-12-04T10:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.911356 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.911433 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.911445 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.911462 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:47 crc kubenswrapper[4943]: I1204 10:16:47.911488 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:47Z","lastTransitionTime":"2025-12-04T10:16:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.013754 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.013795 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.013806 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.013823 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.013833 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:48Z","lastTransitionTime":"2025-12-04T10:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.117409 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.117458 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.117469 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.117486 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.117500 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:48Z","lastTransitionTime":"2025-12-04T10:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.220538 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.220573 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.220582 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.220599 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.220608 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:48Z","lastTransitionTime":"2025-12-04T10:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.322638 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.322685 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.322694 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.322708 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.322717 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:48Z","lastTransitionTime":"2025-12-04T10:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.353874 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/2.log" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.424919 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.424956 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.424965 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.424983 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.424992 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:48Z","lastTransitionTime":"2025-12-04T10:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.527175 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.527224 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.527244 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.527261 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.527273 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:48Z","lastTransitionTime":"2025-12-04T10:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.631070 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.631181 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.631251 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.631278 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.631292 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:48Z","lastTransitionTime":"2025-12-04T10:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.734100 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.734146 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.734158 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.734174 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.734186 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:48Z","lastTransitionTime":"2025-12-04T10:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.835832 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.835867 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.835875 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.835889 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.835898 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:48Z","lastTransitionTime":"2025-12-04T10:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.938609 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.938647 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.938658 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.938675 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:48 crc kubenswrapper[4943]: I1204 10:16:48.938685 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:48Z","lastTransitionTime":"2025-12-04T10:16:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.041623 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.041662 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.041672 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.041685 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.041702 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:49Z","lastTransitionTime":"2025-12-04T10:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.144280 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.144314 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.144322 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.144335 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.144345 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:49Z","lastTransitionTime":"2025-12-04T10:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.246939 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.247285 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.247386 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.247486 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.247597 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:49Z","lastTransitionTime":"2025-12-04T10:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.350663 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.350730 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.350749 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.350781 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.350800 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:49Z","lastTransitionTime":"2025-12-04T10:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.422996 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.423803 4943 scope.go:117] "RemoveContainer" containerID="7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36" Dec 04 10:16:49 crc kubenswrapper[4943]: E1204 10:16:49.423979 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.438102 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.452862 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.452907 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.452919 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.452934 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.452944 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:49Z","lastTransitionTime":"2025-12-04T10:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.454014 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.467075 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.477933 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b54b12fc-f44c-42cf-a647-6ff44c005e5b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07144e00536ba9c9e2f8f9be3199ca05f6b6a864d5ee1fe01b49c4e30a244ae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e348e3b6d23732341818a2d444fc18b715cc7729d49508ba0bec8a349118aba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61720056d847f2e14a4c17ca79e24d8e330530b0ed68c256aa2ddcb35552c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.488927 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.501287 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.511930 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.521804 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.532802 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.549298 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:46Z\\\",\\\"message\\\":\\\"opping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:45.159276 6525 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159565 6525 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159786 6525 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159885 6525 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159901 6525 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159949 6525 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159979 6525 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.160402 6525 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.555745 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.555788 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.555800 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.555821 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.555835 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:49Z","lastTransitionTime":"2025-12-04T10:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.564025 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.564035 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.564122 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.564097 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: E1204 10:16:49.564261 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.564293 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:49 crc kubenswrapper[4943]: E1204 10:16:49.564411 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:49 crc kubenswrapper[4943]: E1204 10:16:49.564424 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:49 crc kubenswrapper[4943]: E1204 10:16:49.564533 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.578472 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.590404 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.601777 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.611697 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.620944 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.632651 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:49Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.658350 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.658387 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.658398 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.658414 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.658424 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:49Z","lastTransitionTime":"2025-12-04T10:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.760705 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.760766 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.760790 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.760812 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.760826 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:49Z","lastTransitionTime":"2025-12-04T10:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.863074 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.863108 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.863126 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.863144 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.863155 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:49Z","lastTransitionTime":"2025-12-04T10:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.965911 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.965955 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.965966 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.965984 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:49 crc kubenswrapper[4943]: I1204 10:16:49.965997 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:49Z","lastTransitionTime":"2025-12-04T10:16:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.068777 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.068848 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.068872 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.068900 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.068921 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:50Z","lastTransitionTime":"2025-12-04T10:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.171718 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.171759 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.171770 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.171785 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.171796 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:50Z","lastTransitionTime":"2025-12-04T10:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.273941 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.273984 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.273995 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.274010 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.274020 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:50Z","lastTransitionTime":"2025-12-04T10:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.376262 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.376295 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.376303 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.376318 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.376331 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:50Z","lastTransitionTime":"2025-12-04T10:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.479528 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.479560 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.479571 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.479587 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.479598 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:50Z","lastTransitionTime":"2025-12-04T10:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.582157 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.582194 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.582243 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.582261 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.582273 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:50Z","lastTransitionTime":"2025-12-04T10:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.684998 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.685121 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.685149 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.685181 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.685243 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:50Z","lastTransitionTime":"2025-12-04T10:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.788770 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.788823 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.788862 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.788879 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.788890 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:50Z","lastTransitionTime":"2025-12-04T10:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.891875 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.891931 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.891942 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.891954 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.891963 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:50Z","lastTransitionTime":"2025-12-04T10:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.994816 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.994855 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.994867 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.994882 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:50 crc kubenswrapper[4943]: I1204 10:16:50.994892 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:50Z","lastTransitionTime":"2025-12-04T10:16:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.097746 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.097792 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.097803 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.097817 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.097827 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:51Z","lastTransitionTime":"2025-12-04T10:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.200515 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.200556 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.200570 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.200585 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.200594 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:51Z","lastTransitionTime":"2025-12-04T10:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.303194 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.303296 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.303316 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.303341 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.303360 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:51Z","lastTransitionTime":"2025-12-04T10:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.405858 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.405915 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.405926 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.405945 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.405964 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:51Z","lastTransitionTime":"2025-12-04T10:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.507965 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.508037 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.508054 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.508105 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.508125 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:51Z","lastTransitionTime":"2025-12-04T10:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.563887 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.564140 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.564300 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:51 crc kubenswrapper[4943]: E1204 10:16:51.564311 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.564542 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:51 crc kubenswrapper[4943]: E1204 10:16:51.564593 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:51 crc kubenswrapper[4943]: E1204 10:16:51.564667 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:51 crc kubenswrapper[4943]: E1204 10:16:51.564735 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.611352 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.611395 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.611404 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.611438 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.611452 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:51Z","lastTransitionTime":"2025-12-04T10:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.714571 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.714631 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.714661 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.714677 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.714686 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:51Z","lastTransitionTime":"2025-12-04T10:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.817232 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.817289 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.817298 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.817311 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.817338 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:51Z","lastTransitionTime":"2025-12-04T10:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.920713 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.920786 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.920806 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.920833 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:51 crc kubenswrapper[4943]: I1204 10:16:51.920860 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:51Z","lastTransitionTime":"2025-12-04T10:16:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.025178 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.025281 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.025303 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.025334 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.025356 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:52Z","lastTransitionTime":"2025-12-04T10:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.128185 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.128273 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.128285 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.128301 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.128312 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:52Z","lastTransitionTime":"2025-12-04T10:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.231071 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.231119 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.231131 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.231151 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.231163 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:52Z","lastTransitionTime":"2025-12-04T10:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.334271 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.334345 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.334358 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.334378 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.334390 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:52Z","lastTransitionTime":"2025-12-04T10:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.437739 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.438553 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.438716 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.438848 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.438989 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:52Z","lastTransitionTime":"2025-12-04T10:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.541782 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.542109 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.542189 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.542281 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.542352 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:52Z","lastTransitionTime":"2025-12-04T10:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.586845 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.603221 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b54b12fc-f44c-42cf-a647-6ff44c005e5b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07144e00536ba9c9e2f8f9be3199ca05f6b6a864d5ee1fe01b49c4e30a244ae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e348e3b6d23732341818a2d444fc18b715cc7729d49508ba0bec8a349118aba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61720056d847f2e14a4c17ca79e24d8e330530b0ed68c256aa2ddcb35552c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.619867 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.642093 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.645630 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.645666 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.645674 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.645689 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.645698 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:52Z","lastTransitionTime":"2025-12-04T10:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.657464 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.669088 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.681782 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.703338 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:46Z\\\",\\\"message\\\":\\\"opping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:45.159276 6525 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159565 6525 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159786 6525 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159885 6525 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159901 6525 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159949 6525 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159979 6525 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.160402 6525 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.720572 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.730381 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.741290 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.748455 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.748507 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.748516 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.748530 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.748540 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:52Z","lastTransitionTime":"2025-12-04T10:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.753314 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.768395 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.783399 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.798489 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.809717 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.823796 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:52Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.851304 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.851714 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.851861 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.851939 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.852017 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:52Z","lastTransitionTime":"2025-12-04T10:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.954972 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.955014 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.955025 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.955041 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:52 crc kubenswrapper[4943]: I1204 10:16:52.955050 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:52Z","lastTransitionTime":"2025-12-04T10:16:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.057712 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.057744 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.057753 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.057766 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.057775 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:53Z","lastTransitionTime":"2025-12-04T10:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.160118 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.160157 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.160167 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.160182 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.160193 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:53Z","lastTransitionTime":"2025-12-04T10:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.262395 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.262422 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.262436 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.262448 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.262459 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:53Z","lastTransitionTime":"2025-12-04T10:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.365579 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.365618 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.365626 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.365641 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.365652 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:53Z","lastTransitionTime":"2025-12-04T10:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.467347 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.467392 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.467403 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.467418 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.467428 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:53Z","lastTransitionTime":"2025-12-04T10:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.564326 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.564423 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:53 crc kubenswrapper[4943]: E1204 10:16:53.564462 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.564340 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.564522 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:53 crc kubenswrapper[4943]: E1204 10:16:53.564583 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:53 crc kubenswrapper[4943]: E1204 10:16:53.564756 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:53 crc kubenswrapper[4943]: E1204 10:16:53.564818 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.571311 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.571350 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.571360 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.571375 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.571385 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:53Z","lastTransitionTime":"2025-12-04T10:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.673777 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.673818 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.673826 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.673840 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.673851 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:53Z","lastTransitionTime":"2025-12-04T10:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.776769 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.776825 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.776843 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.776864 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.776879 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:53Z","lastTransitionTime":"2025-12-04T10:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.879299 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.879360 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.879369 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.879384 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.879392 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:53Z","lastTransitionTime":"2025-12-04T10:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.981539 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.981575 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.981584 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.981599 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:53 crc kubenswrapper[4943]: I1204 10:16:53.981608 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:53Z","lastTransitionTime":"2025-12-04T10:16:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.083427 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.083467 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.083477 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.083513 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.083530 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.186765 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.186812 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.186830 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.186848 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.186863 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.288896 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.288927 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.288935 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.288947 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.288955 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.391606 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.391641 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.391652 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.391667 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.391677 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.493575 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.493674 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.493692 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.493716 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.493733 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.596118 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.596157 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.596165 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.596178 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.596186 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.698860 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.698927 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.698938 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.698952 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.698960 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.728383 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.728432 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.728446 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.728462 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.728472 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: E1204 10:16:54.762699 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:54Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.770543 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.770595 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.770609 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.770626 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.770638 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: E1204 10:16:54.789823 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:54Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.794442 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.794498 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.794512 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.794530 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.794542 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: E1204 10:16:54.814446 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:54Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.818579 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.818619 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.818627 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.818647 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.818656 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: E1204 10:16:54.832334 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:54Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.836288 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.836330 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.836344 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.836362 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.836372 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: E1204 10:16:54.852288 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:54Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:16:54Z is after 2025-08-24T17:21:41Z" Dec 04 10:16:54 crc kubenswrapper[4943]: E1204 10:16:54.852450 4943 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.855008 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.855036 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.855046 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.855063 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.855073 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.957339 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.957375 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.957384 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.957399 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:54 crc kubenswrapper[4943]: I1204 10:16:54.957408 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:54Z","lastTransitionTime":"2025-12-04T10:16:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.060077 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.060121 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.060130 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.060146 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.060155 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:55Z","lastTransitionTime":"2025-12-04T10:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.162559 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.162940 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.162953 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.163014 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.163029 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:55Z","lastTransitionTime":"2025-12-04T10:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.265869 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.265922 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.265933 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.265959 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.265974 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:55Z","lastTransitionTime":"2025-12-04T10:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.368629 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.368672 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.368683 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.368699 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.368710 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:55Z","lastTransitionTime":"2025-12-04T10:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.471401 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.471475 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.471493 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.471511 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.471523 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:55Z","lastTransitionTime":"2025-12-04T10:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.563863 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.563961 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:55 crc kubenswrapper[4943]: E1204 10:16:55.564009 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.563863 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.563880 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:55 crc kubenswrapper[4943]: E1204 10:16:55.564118 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:55 crc kubenswrapper[4943]: E1204 10:16:55.564191 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:55 crc kubenswrapper[4943]: E1204 10:16:55.564291 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.574246 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.574288 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.574300 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.574314 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.574324 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:55Z","lastTransitionTime":"2025-12-04T10:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.676770 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.676812 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.676820 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.676836 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.676845 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:55Z","lastTransitionTime":"2025-12-04T10:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.779685 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.779728 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.779740 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.779762 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.779776 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:55Z","lastTransitionTime":"2025-12-04T10:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.882726 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.882779 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.882794 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.882819 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.882836 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:55Z","lastTransitionTime":"2025-12-04T10:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.985507 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.985540 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.985549 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.985566 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:55 crc kubenswrapper[4943]: I1204 10:16:55.985576 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:55Z","lastTransitionTime":"2025-12-04T10:16:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.087865 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.087910 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.087919 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.087936 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.087947 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:56Z","lastTransitionTime":"2025-12-04T10:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.190888 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.190923 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.190933 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.190949 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.190959 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:56Z","lastTransitionTime":"2025-12-04T10:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.292758 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.292797 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.292808 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.292828 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.292841 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:56Z","lastTransitionTime":"2025-12-04T10:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.394831 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.394893 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.394902 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.394916 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.394929 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:56Z","lastTransitionTime":"2025-12-04T10:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.498074 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.498112 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.498123 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.498138 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.498150 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:56Z","lastTransitionTime":"2025-12-04T10:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.600245 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.600282 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.600293 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.600307 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.600319 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:56Z","lastTransitionTime":"2025-12-04T10:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.702539 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.702598 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.702615 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.702641 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.702657 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:56Z","lastTransitionTime":"2025-12-04T10:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.805231 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.805269 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.805279 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.805292 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.805306 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:56Z","lastTransitionTime":"2025-12-04T10:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.907545 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.907599 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.907608 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.907621 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:56 crc kubenswrapper[4943]: I1204 10:16:56.907630 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:56Z","lastTransitionTime":"2025-12-04T10:16:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.010114 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.010158 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.010168 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.010182 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.010193 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:57Z","lastTransitionTime":"2025-12-04T10:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.112969 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.113010 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.113019 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.113033 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.113046 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:57Z","lastTransitionTime":"2025-12-04T10:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.216915 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.216995 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.217006 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.217022 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.217038 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:57Z","lastTransitionTime":"2025-12-04T10:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.319792 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.319831 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.319840 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.319854 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.319864 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:57Z","lastTransitionTime":"2025-12-04T10:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.422999 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.423036 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.423047 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.423063 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.423073 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:57Z","lastTransitionTime":"2025-12-04T10:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.525814 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.525881 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.525892 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.525927 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.525940 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:57Z","lastTransitionTime":"2025-12-04T10:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.564575 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.564625 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.564590 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.564575 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:57 crc kubenswrapper[4943]: E1204 10:16:57.564718 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:57 crc kubenswrapper[4943]: E1204 10:16:57.564793 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:57 crc kubenswrapper[4943]: E1204 10:16:57.564871 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:57 crc kubenswrapper[4943]: E1204 10:16:57.564948 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.628586 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.628635 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.628645 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.628659 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.628668 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:57Z","lastTransitionTime":"2025-12-04T10:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.730815 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.730876 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.730886 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.730903 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.730912 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:57Z","lastTransitionTime":"2025-12-04T10:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.833640 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.833699 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.833710 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.833729 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.833740 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:57Z","lastTransitionTime":"2025-12-04T10:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.936551 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.936597 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.936609 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.936625 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:57 crc kubenswrapper[4943]: I1204 10:16:57.936639 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:57Z","lastTransitionTime":"2025-12-04T10:16:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.038990 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.039044 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.039058 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.039074 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.039084 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:58Z","lastTransitionTime":"2025-12-04T10:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.142285 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.142439 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.142451 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.142469 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.142480 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:58Z","lastTransitionTime":"2025-12-04T10:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.245721 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.245784 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.245795 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.245810 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.245821 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:58Z","lastTransitionTime":"2025-12-04T10:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.348634 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.348680 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.348692 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.348710 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.348722 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:58Z","lastTransitionTime":"2025-12-04T10:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.451791 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.451839 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.451851 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.451868 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.451879 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:58Z","lastTransitionTime":"2025-12-04T10:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.554366 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.554430 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.554440 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.554457 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.554469 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:58Z","lastTransitionTime":"2025-12-04T10:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.657537 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.657575 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.657584 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.657611 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.657622 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:58Z","lastTransitionTime":"2025-12-04T10:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.759692 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.759750 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.759763 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.759779 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.759789 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:58Z","lastTransitionTime":"2025-12-04T10:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.863013 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.863059 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.863067 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.863082 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.863091 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:58Z","lastTransitionTime":"2025-12-04T10:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.965754 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.965832 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.965848 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.965866 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:58 crc kubenswrapper[4943]: I1204 10:16:58.965877 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:58Z","lastTransitionTime":"2025-12-04T10:16:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.068182 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.068244 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.068256 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.068272 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.068286 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:59Z","lastTransitionTime":"2025-12-04T10:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.170984 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.171040 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.171054 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.171072 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.171086 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:59Z","lastTransitionTime":"2025-12-04T10:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.273432 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.273491 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.273508 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.273536 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.273550 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:59Z","lastTransitionTime":"2025-12-04T10:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.375916 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.375996 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.376020 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.376047 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.376074 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:59Z","lastTransitionTime":"2025-12-04T10:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.479270 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.479334 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.479349 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.479371 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.479385 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:59Z","lastTransitionTime":"2025-12-04T10:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.564452 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.564512 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.564556 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.564450 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:16:59 crc kubenswrapper[4943]: E1204 10:16:59.564604 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:16:59 crc kubenswrapper[4943]: E1204 10:16:59.564673 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:16:59 crc kubenswrapper[4943]: E1204 10:16:59.564725 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:16:59 crc kubenswrapper[4943]: E1204 10:16:59.564790 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.582152 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.582217 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.582239 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.582255 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.582266 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:59Z","lastTransitionTime":"2025-12-04T10:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.684476 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.684512 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.684521 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.684533 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.684543 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:59Z","lastTransitionTime":"2025-12-04T10:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.787748 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.787815 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.787840 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.787870 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.787894 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:59Z","lastTransitionTime":"2025-12-04T10:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.890639 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.890686 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.890694 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.890709 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.890718 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:59Z","lastTransitionTime":"2025-12-04T10:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.993412 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.993452 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.993461 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.993476 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:16:59 crc kubenswrapper[4943]: I1204 10:16:59.993485 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:16:59Z","lastTransitionTime":"2025-12-04T10:16:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.095748 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.095805 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.095820 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.095841 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.095860 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:00Z","lastTransitionTime":"2025-12-04T10:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.198750 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.198784 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.198794 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.198809 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.198819 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:00Z","lastTransitionTime":"2025-12-04T10:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.226384 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:00 crc kubenswrapper[4943]: E1204 10:17:00.226563 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:17:00 crc kubenswrapper[4943]: E1204 10:17:00.226713 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs podName:eb85a8e0-6c90-47cb-a1b5-9ecd5244c710 nodeName:}" failed. No retries permitted until 2025-12-04 10:17:32.226674695 +0000 UTC m=+100.815650563 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs") pod "network-metrics-daemon-hw5pj" (UID: "eb85a8e0-6c90-47cb-a1b5-9ecd5244c710") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.301936 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.301993 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.302010 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.302026 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.302037 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:00Z","lastTransitionTime":"2025-12-04T10:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.404861 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.404919 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.404931 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.404959 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.404972 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:00Z","lastTransitionTime":"2025-12-04T10:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.507680 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.507721 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.507730 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.507744 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.507753 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:00Z","lastTransitionTime":"2025-12-04T10:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.609834 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.609878 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.609886 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.609904 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.609913 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:00Z","lastTransitionTime":"2025-12-04T10:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.712139 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.712179 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.712189 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.712232 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.712244 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:00Z","lastTransitionTime":"2025-12-04T10:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.814363 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.814398 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.814406 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.814420 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.814429 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:00Z","lastTransitionTime":"2025-12-04T10:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.917350 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.917395 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.917407 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.917424 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:00 crc kubenswrapper[4943]: I1204 10:17:00.917435 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:00Z","lastTransitionTime":"2025-12-04T10:17:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.020144 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.020200 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.020238 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.020256 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.020265 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:01Z","lastTransitionTime":"2025-12-04T10:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.122555 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.122603 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.122614 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.122630 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.122641 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:01Z","lastTransitionTime":"2025-12-04T10:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.226256 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.226342 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.226369 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.226396 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.226413 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:01Z","lastTransitionTime":"2025-12-04T10:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.328725 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.328774 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.328784 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.328803 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.328813 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:01Z","lastTransitionTime":"2025-12-04T10:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.431163 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.431243 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.431258 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.431278 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.431288 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:01Z","lastTransitionTime":"2025-12-04T10:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.533344 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.533387 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.533399 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.533417 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.533429 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:01Z","lastTransitionTime":"2025-12-04T10:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.563710 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.563711 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.563753 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.563815 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:01 crc kubenswrapper[4943]: E1204 10:17:01.563964 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:01 crc kubenswrapper[4943]: E1204 10:17:01.564064 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:01 crc kubenswrapper[4943]: E1204 10:17:01.564165 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:01 crc kubenswrapper[4943]: E1204 10:17:01.564288 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.636166 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.636211 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.636242 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.636259 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.636271 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:01Z","lastTransitionTime":"2025-12-04T10:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.738874 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.738926 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.738941 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.738959 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.738971 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:01Z","lastTransitionTime":"2025-12-04T10:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.841141 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.841217 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.841231 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.841250 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.841261 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:01Z","lastTransitionTime":"2025-12-04T10:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.944017 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.944065 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.944077 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.944094 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:01 crc kubenswrapper[4943]: I1204 10:17:01.944107 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:01Z","lastTransitionTime":"2025-12-04T10:17:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.046453 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.046511 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.046522 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.046541 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.046595 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:02Z","lastTransitionTime":"2025-12-04T10:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.148954 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.149027 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.149038 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.149056 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.149068 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:02Z","lastTransitionTime":"2025-12-04T10:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.302111 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.302151 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.302161 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.302178 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.302187 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:02Z","lastTransitionTime":"2025-12-04T10:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.405303 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.405346 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.405357 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.405396 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.405411 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:02Z","lastTransitionTime":"2025-12-04T10:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.508628 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.508994 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.509062 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.509184 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.509322 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:02Z","lastTransitionTime":"2025-12-04T10:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.565572 4943 scope.go:117] "RemoveContainer" containerID="7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36" Dec 04 10:17:02 crc kubenswrapper[4943]: E1204 10:17:02.566051 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.581291 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.595357 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.611303 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.611600 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.611628 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.611636 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.611654 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.611663 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:02Z","lastTransitionTime":"2025-12-04T10:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.629437 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b54b12fc-f44c-42cf-a647-6ff44c005e5b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07144e00536ba9c9e2f8f9be3199ca05f6b6a864d5ee1fe01b49c4e30a244ae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e348e3b6d23732341818a2d444fc18b715cc7729d49508ba0bec8a349118aba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61720056d847f2e14a4c17ca79e24d8e330530b0ed68c256aa2ddcb35552c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.650656 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.672633 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.688011 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.703984 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.715230 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.715282 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.715293 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.715315 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.715328 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:02Z","lastTransitionTime":"2025-12-04T10:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.721362 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.742131 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:46Z\\\",\\\"message\\\":\\\"opping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:45.159276 6525 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159565 6525 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159786 6525 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159885 6525 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159901 6525 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159949 6525 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159979 6525 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.160402 6525 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.761367 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.773970 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.788215 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.801888 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.819274 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.819320 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.819330 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.819349 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.819363 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:02Z","lastTransitionTime":"2025-12-04T10:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.822973 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.842646 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.856810 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:02Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.927086 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.927170 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.927180 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.927199 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:02 crc kubenswrapper[4943]: I1204 10:17:02.927228 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:02Z","lastTransitionTime":"2025-12-04T10:17:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.030547 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.033935 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.033949 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.033964 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.033986 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:03Z","lastTransitionTime":"2025-12-04T10:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.136605 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.136657 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.136670 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.136691 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.136704 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:03Z","lastTransitionTime":"2025-12-04T10:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.240518 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.240569 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.240610 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.240644 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.240666 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:03Z","lastTransitionTime":"2025-12-04T10:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.343249 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.343282 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.343296 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.343309 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.343318 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:03Z","lastTransitionTime":"2025-12-04T10:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.407088 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wkbf5_b9cb494f-7d4c-45f2-8b9b-e35c42b41c79/kube-multus/0.log" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.407223 4943 generic.go:334] "Generic (PLEG): container finished" podID="b9cb494f-7d4c-45f2-8b9b-e35c42b41c79" containerID="babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793" exitCode=1 Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.407280 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wkbf5" event={"ID":"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79","Type":"ContainerDied","Data":"babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793"} Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.408205 4943 scope.go:117] "RemoveContainer" containerID="babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.424469 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.439050 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b54b12fc-f44c-42cf-a647-6ff44c005e5b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07144e00536ba9c9e2f8f9be3199ca05f6b6a864d5ee1fe01b49c4e30a244ae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e348e3b6d23732341818a2d444fc18b715cc7729d49508ba0bec8a349118aba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61720056d847f2e14a4c17ca79e24d8e330530b0ed68c256aa2ddcb35552c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.445544 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.445589 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.445600 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.445617 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.445629 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:03Z","lastTransitionTime":"2025-12-04T10:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.454148 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.473423 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.486218 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.499582 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.513288 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:17:02Z\\\",\\\"message\\\":\\\"2025-12-04T10:16:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_69e1ca66-d180-4c54-93e0-dc534d876e60\\\\n2025-12-04T10:16:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_69e1ca66-d180-4c54-93e0-dc534d876e60 to /host/opt/cni/bin/\\\\n2025-12-04T10:16:17Z [verbose] multus-daemon started\\\\n2025-12-04T10:16:17Z [verbose] Readiness Indicator file check\\\\n2025-12-04T10:17:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.530877 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:46Z\\\",\\\"message\\\":\\\"opping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:45.159276 6525 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159565 6525 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159786 6525 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159885 6525 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159901 6525 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159949 6525 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159979 6525 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.160402 6525 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.543700 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.547583 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.547633 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.547643 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.547657 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.547692 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:03Z","lastTransitionTime":"2025-12-04T10:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.554422 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.564672 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.564740 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.565243 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.565414 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:03 crc kubenswrapper[4943]: E1204 10:17:03.565554 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.565443 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: E1204 10:17:03.565401 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:03 crc kubenswrapper[4943]: E1204 10:17:03.565496 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:03 crc kubenswrapper[4943]: E1204 10:17:03.566019 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.579264 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.593706 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.606741 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.624051 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.635588 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.645721 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:03Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.650472 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.650514 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.650524 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.650539 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.650549 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:03Z","lastTransitionTime":"2025-12-04T10:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.753394 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.754335 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.754430 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.754503 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.754565 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:03Z","lastTransitionTime":"2025-12-04T10:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.856927 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.857201 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.857321 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.857413 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.857507 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:03Z","lastTransitionTime":"2025-12-04T10:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.961115 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.961457 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.961583 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.961720 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:03 crc kubenswrapper[4943]: I1204 10:17:03.961850 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:03Z","lastTransitionTime":"2025-12-04T10:17:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.065868 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.065921 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.065939 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.065960 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.065975 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:04Z","lastTransitionTime":"2025-12-04T10:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.168786 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.168827 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.168836 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.168855 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.168864 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:04Z","lastTransitionTime":"2025-12-04T10:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.271065 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.271144 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.271157 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.271175 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.271188 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:04Z","lastTransitionTime":"2025-12-04T10:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.372913 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.372952 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.372963 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.372979 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.372991 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:04Z","lastTransitionTime":"2025-12-04T10:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.415061 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wkbf5_b9cb494f-7d4c-45f2-8b9b-e35c42b41c79/kube-multus/0.log" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.415125 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wkbf5" event={"ID":"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79","Type":"ContainerStarted","Data":"635d7d730273976acb953de2147ab4353647915f5f9898a96c04558ab8262a1d"} Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.430176 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.441778 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.453929 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.463942 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b54b12fc-f44c-42cf-a647-6ff44c005e5b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07144e00536ba9c9e2f8f9be3199ca05f6b6a864d5ee1fe01b49c4e30a244ae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e348e3b6d23732341818a2d444fc18b715cc7729d49508ba0bec8a349118aba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61720056d847f2e14a4c17ca79e24d8e330530b0ed68c256aa2ddcb35552c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.477019 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.477079 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.477093 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.477113 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.477132 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:04Z","lastTransitionTime":"2025-12-04T10:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.482544 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.496958 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.511600 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.523941 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.537789 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635d7d730273976acb953de2147ab4353647915f5f9898a96c04558ab8262a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:17:02Z\\\",\\\"message\\\":\\\"2025-12-04T10:16:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_69e1ca66-d180-4c54-93e0-dc534d876e60\\\\n2025-12-04T10:16:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_69e1ca66-d180-4c54-93e0-dc534d876e60 to /host/opt/cni/bin/\\\\n2025-12-04T10:16:17Z [verbose] multus-daemon started\\\\n2025-12-04T10:16:17Z [verbose] Readiness Indicator file check\\\\n2025-12-04T10:17:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:17:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.558787 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:46Z\\\",\\\"message\\\":\\\"opping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:45.159276 6525 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159565 6525 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159786 6525 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159885 6525 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159901 6525 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159949 6525 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159979 6525 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.160402 6525 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.570920 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.579388 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.579432 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.579443 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.579460 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.579471 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:04Z","lastTransitionTime":"2025-12-04T10:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.583412 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.597110 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.608840 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.620394 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.629915 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.639749 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.682460 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.682505 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.682520 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.682537 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.682551 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:04Z","lastTransitionTime":"2025-12-04T10:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.784752 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.784820 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.784832 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.784849 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.784861 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:04Z","lastTransitionTime":"2025-12-04T10:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.887963 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.888019 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.888033 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.888050 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.888063 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:04Z","lastTransitionTime":"2025-12-04T10:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.976649 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.976694 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.976707 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.976723 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.976737 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:04Z","lastTransitionTime":"2025-12-04T10:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:04 crc kubenswrapper[4943]: E1204 10:17:04.989771 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:04Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.992868 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.992893 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.992901 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.992915 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:04 crc kubenswrapper[4943]: I1204 10:17:04.992925 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:04Z","lastTransitionTime":"2025-12-04T10:17:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: E1204 10:17:05.005312 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:05Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.009095 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.009125 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.009133 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.009145 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.009154 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: E1204 10:17:05.022411 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:05Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.026163 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.026193 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.026218 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.026233 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.026244 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: E1204 10:17:05.037161 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:05Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.041334 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.041367 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.041376 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.041386 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.041409 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: E1204 10:17:05.053292 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:05Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:05 crc kubenswrapper[4943]: E1204 10:17:05.053425 4943 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.054872 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.054904 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.054916 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.054933 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.054944 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.157832 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.158200 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.158308 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.158419 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.158501 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.260564 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.260611 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.260632 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.260651 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.260667 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.363411 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.363464 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.363476 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.363495 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.363508 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.466048 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.466333 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.466457 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.466797 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.466895 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.564022 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.564190 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.564041 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.564022 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:05 crc kubenswrapper[4943]: E1204 10:17:05.564601 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:05 crc kubenswrapper[4943]: E1204 10:17:05.564822 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:05 crc kubenswrapper[4943]: E1204 10:17:05.564924 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:05 crc kubenswrapper[4943]: E1204 10:17:05.565007 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.569705 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.569855 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.569936 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.570022 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.570103 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.673232 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.673272 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.673283 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.673298 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.673325 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.775736 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.775779 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.775791 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.775807 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.775819 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.879356 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.879670 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.879680 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.879698 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.879707 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.982309 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.982374 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.982388 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.982404 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:05 crc kubenswrapper[4943]: I1204 10:17:05.982416 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:05Z","lastTransitionTime":"2025-12-04T10:17:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.085317 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.085371 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.085388 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.085405 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.085417 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:06Z","lastTransitionTime":"2025-12-04T10:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.188266 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.188353 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.188375 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.188404 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.188425 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:06Z","lastTransitionTime":"2025-12-04T10:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.291826 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.291884 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.291906 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.291925 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.291937 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:06Z","lastTransitionTime":"2025-12-04T10:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.394835 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.394877 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.394888 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.394903 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.394913 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:06Z","lastTransitionTime":"2025-12-04T10:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.496890 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.496948 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.496960 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.496975 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.496986 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:06Z","lastTransitionTime":"2025-12-04T10:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.599360 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.599404 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.599413 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.599426 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.599437 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:06Z","lastTransitionTime":"2025-12-04T10:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.702006 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.702076 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.702091 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.702111 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.702123 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:06Z","lastTransitionTime":"2025-12-04T10:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.805263 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.805303 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.805312 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.805326 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.805336 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:06Z","lastTransitionTime":"2025-12-04T10:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.907947 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.908027 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.908043 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.908072 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:06 crc kubenswrapper[4943]: I1204 10:17:06.908085 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:06Z","lastTransitionTime":"2025-12-04T10:17:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.012567 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.012629 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.012642 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.012666 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.012679 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:07Z","lastTransitionTime":"2025-12-04T10:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.115804 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.115862 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.115893 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.115912 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.115924 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:07Z","lastTransitionTime":"2025-12-04T10:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.219185 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.219272 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.219291 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.219315 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.219332 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:07Z","lastTransitionTime":"2025-12-04T10:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.322288 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.322334 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.322344 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.322360 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.322370 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:07Z","lastTransitionTime":"2025-12-04T10:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.424799 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.424856 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.424869 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.424885 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.424894 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:07Z","lastTransitionTime":"2025-12-04T10:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.527595 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.527647 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.527667 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.527696 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.527715 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:07Z","lastTransitionTime":"2025-12-04T10:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.564556 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.564610 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.564643 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.564636 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:07 crc kubenswrapper[4943]: E1204 10:17:07.564772 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:07 crc kubenswrapper[4943]: E1204 10:17:07.564923 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:07 crc kubenswrapper[4943]: E1204 10:17:07.565058 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:07 crc kubenswrapper[4943]: E1204 10:17:07.565172 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.630417 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.630463 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.630474 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.630492 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.630506 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:07Z","lastTransitionTime":"2025-12-04T10:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.733221 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.733258 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.733269 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.733284 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.733296 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:07Z","lastTransitionTime":"2025-12-04T10:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.835442 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.835482 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.835495 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.835510 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.835522 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:07Z","lastTransitionTime":"2025-12-04T10:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.938401 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.938451 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.938462 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.938480 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:07 crc kubenswrapper[4943]: I1204 10:17:07.938491 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:07Z","lastTransitionTime":"2025-12-04T10:17:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.040632 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.040689 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.040700 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.040721 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.040733 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:08Z","lastTransitionTime":"2025-12-04T10:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.144156 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.144194 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.144221 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.144237 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.144247 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:08Z","lastTransitionTime":"2025-12-04T10:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.246953 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.246991 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.247002 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.247018 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.247029 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:08Z","lastTransitionTime":"2025-12-04T10:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.349464 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.349507 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.349544 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.349562 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.349572 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:08Z","lastTransitionTime":"2025-12-04T10:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.453175 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.453257 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.453270 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.453289 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.453305 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:08Z","lastTransitionTime":"2025-12-04T10:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.556052 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.556106 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.556118 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.556135 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.556149 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:08Z","lastTransitionTime":"2025-12-04T10:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.658435 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.658489 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.658500 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.658515 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.658526 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:08Z","lastTransitionTime":"2025-12-04T10:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.760291 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.760335 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.760347 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.760365 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.760376 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:08Z","lastTransitionTime":"2025-12-04T10:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.862329 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.862384 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.862399 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.862417 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.862430 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:08Z","lastTransitionTime":"2025-12-04T10:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.964766 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.964840 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.964854 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.964877 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:08 crc kubenswrapper[4943]: I1204 10:17:08.964898 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:08Z","lastTransitionTime":"2025-12-04T10:17:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.067356 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.067400 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.067409 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.067424 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.067435 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:09Z","lastTransitionTime":"2025-12-04T10:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.169483 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.169525 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.169536 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.169552 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.169563 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:09Z","lastTransitionTime":"2025-12-04T10:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.271986 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.272028 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.272043 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.272060 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.272073 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:09Z","lastTransitionTime":"2025-12-04T10:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.374905 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.374958 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.374970 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.374987 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.375000 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:09Z","lastTransitionTime":"2025-12-04T10:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.477739 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.477768 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.477776 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.477789 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.477797 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:09Z","lastTransitionTime":"2025-12-04T10:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.564053 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.564088 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.564125 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.564127 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:09 crc kubenswrapper[4943]: E1204 10:17:09.564216 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:09 crc kubenswrapper[4943]: E1204 10:17:09.564268 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:09 crc kubenswrapper[4943]: E1204 10:17:09.564328 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:09 crc kubenswrapper[4943]: E1204 10:17:09.564439 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.579599 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.579626 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.579635 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.579647 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.579655 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:09Z","lastTransitionTime":"2025-12-04T10:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.682379 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.682421 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.682433 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.682450 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.682462 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:09Z","lastTransitionTime":"2025-12-04T10:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.784891 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.784937 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.784954 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.784970 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.784980 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:09Z","lastTransitionTime":"2025-12-04T10:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.887437 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.887476 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.887487 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.887502 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.887517 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:09Z","lastTransitionTime":"2025-12-04T10:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.989806 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.989854 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.989867 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.989885 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:09 crc kubenswrapper[4943]: I1204 10:17:09.989896 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:09Z","lastTransitionTime":"2025-12-04T10:17:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.091706 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.091760 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.091769 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.091785 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.091798 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:10Z","lastTransitionTime":"2025-12-04T10:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.193901 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.193952 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.193964 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.193986 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.194003 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:10Z","lastTransitionTime":"2025-12-04T10:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.296833 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.296902 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.296914 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.296930 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.296941 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:10Z","lastTransitionTime":"2025-12-04T10:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.400263 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.400326 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.400339 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.400359 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.400374 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:10Z","lastTransitionTime":"2025-12-04T10:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.502247 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.502295 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.502304 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.502324 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.502334 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:10Z","lastTransitionTime":"2025-12-04T10:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.604990 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.605032 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.605040 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.605057 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.605066 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:10Z","lastTransitionTime":"2025-12-04T10:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.707874 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.707922 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.707933 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.707951 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.707962 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:10Z","lastTransitionTime":"2025-12-04T10:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.810223 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.810276 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.810288 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.810306 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.810317 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:10Z","lastTransitionTime":"2025-12-04T10:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.913312 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.913818 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.913911 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.913998 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:10 crc kubenswrapper[4943]: I1204 10:17:10.914094 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:10Z","lastTransitionTime":"2025-12-04T10:17:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.016264 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.016320 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.016333 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.016353 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.016365 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:11Z","lastTransitionTime":"2025-12-04T10:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.120223 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.120268 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.120278 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.120293 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.120302 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:11Z","lastTransitionTime":"2025-12-04T10:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.222428 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.222475 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.222484 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.222499 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.222508 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:11Z","lastTransitionTime":"2025-12-04T10:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.324931 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.324980 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.324991 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.325007 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.325020 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:11Z","lastTransitionTime":"2025-12-04T10:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.426917 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.426961 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.426975 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.426992 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.427003 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:11Z","lastTransitionTime":"2025-12-04T10:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.529347 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.529411 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.529500 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.529569 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.529586 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:11Z","lastTransitionTime":"2025-12-04T10:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.564057 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.564085 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.564091 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.564071 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:11 crc kubenswrapper[4943]: E1204 10:17:11.564188 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:11 crc kubenswrapper[4943]: E1204 10:17:11.564266 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:11 crc kubenswrapper[4943]: E1204 10:17:11.564370 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:11 crc kubenswrapper[4943]: E1204 10:17:11.564526 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.631772 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.631821 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.631838 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.631860 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.631877 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:11Z","lastTransitionTime":"2025-12-04T10:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.734446 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.734496 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.734535 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.734552 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.734563 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:11Z","lastTransitionTime":"2025-12-04T10:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.836918 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.836969 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.836982 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.836998 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.837011 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:11Z","lastTransitionTime":"2025-12-04T10:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.939273 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.939332 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.939341 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.939356 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:11 crc kubenswrapper[4943]: I1204 10:17:11.939365 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:11Z","lastTransitionTime":"2025-12-04T10:17:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.041226 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.041277 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.041288 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.041304 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.041318 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:12Z","lastTransitionTime":"2025-12-04T10:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.144484 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.144557 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.144580 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.144612 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.144638 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:12Z","lastTransitionTime":"2025-12-04T10:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.247185 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.247253 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.247262 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.247276 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.247285 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:12Z","lastTransitionTime":"2025-12-04T10:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.348920 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.348967 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.348982 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.349001 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.349013 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:12Z","lastTransitionTime":"2025-12-04T10:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.451620 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.451663 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.451672 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.451689 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.451702 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:12Z","lastTransitionTime":"2025-12-04T10:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.554378 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.554431 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.554440 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.554458 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.554468 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:12Z","lastTransitionTime":"2025-12-04T10:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.580922 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.593997 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b54b12fc-f44c-42cf-a647-6ff44c005e5b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07144e00536ba9c9e2f8f9be3199ca05f6b6a864d5ee1fe01b49c4e30a244ae1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0e348e3b6d23732341818a2d444fc18b715cc7729d49508ba0bec8a349118aba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a61720056d847f2e14a4c17ca79e24d8e330530b0ed68c256aa2ddcb35552c25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://922647c9ef27f14df3dd6fae1b6cd3a2d65a5835a05fbded8444f917b4d5701b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.607305 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5b4ce278dead21eee525376d05b4ff6ff8afdc299b25f73cde398ba0c0964f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80613851f81f580a5722d3381dae8ed6452b51c3635f0d569edab2b31edd3de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.622095 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"28b3b2d6-31eb-421d-80f5-71304da27696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3095264e281b94f0466df42c03a899b325dc534100cdb37fc28b6601b971803c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e62e7b943c3797f5926ec6c847016b1ca879d4c849d5fd11c6fa687400e0cca8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://025c66890705e04f2607a624f62e83ffb0ffe50b600654d71701461f8380a032\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0030c3725534d464e635a2c73273f483cdb1cc502af73dd71911dcdfd4b6e223\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00af85359afad990fc0e752e0a12901559eba05692b9f6cbcaa1bd6f00dc020d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc956880807dfdaad1203c707c36983c7732c7adf1047d0b0edbe5326a331a3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://29f479e4918bc708a11d946cf6171b2d968be0ec7cf79450c679c99fde634667\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sxs29\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9nxvx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.634316 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6c68ccf298ac3d49048b4f259beacd93da1e867a78d9fd55322e7d5c32a4987\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.647256 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.655968 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.656023 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.656035 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.656052 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.656065 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:12Z","lastTransitionTime":"2025-12-04T10:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.664532 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-wkbf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://635d7d730273976acb953de2147ab4353647915f5f9898a96c04558ab8262a1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:17:02Z\\\",\\\"message\\\":\\\"2025-12-04T10:16:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_69e1ca66-d180-4c54-93e0-dc534d876e60\\\\n2025-12-04T10:16:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_69e1ca66-d180-4c54-93e0-dc534d876e60 to /host/opt/cni/bin/\\\\n2025-12-04T10:16:17Z [verbose] multus-daemon started\\\\n2025-12-04T10:16:17Z [verbose] Readiness Indicator file check\\\\n2025-12-04T10:17:02Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:17:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w667f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-multus\"/\"multus-wkbf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.683423 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0498f58-c99b-4fa8-a404-f7f43199c29f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T10:16:46Z\\\",\\\"message\\\":\\\"opping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1204 10:16:45.159276 6525 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159565 6525 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159786 6525 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159885 6525 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.159901 6525 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159949 6525 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 10:16:45.159979 6525 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 10:16:45.160402 6525 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7wh9m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-7g4wf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.693185 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.702446 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.712983 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.732653 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.747073 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.758625 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.758663 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.758674 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.758690 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.758701 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:12Z","lastTransitionTime":"2025-12-04T10:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.761357 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.773217 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.783780 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.795581 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:12Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.861351 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.861402 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.861414 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.861432 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.861446 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:12Z","lastTransitionTime":"2025-12-04T10:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.964065 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.964333 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.964436 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.964531 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:12 crc kubenswrapper[4943]: I1204 10:17:12.964611 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:12Z","lastTransitionTime":"2025-12-04T10:17:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.067816 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.067879 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.067896 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.067921 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.067938 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:13Z","lastTransitionTime":"2025-12-04T10:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.170763 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.170805 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.170817 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.170832 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.170842 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:13Z","lastTransitionTime":"2025-12-04T10:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.273812 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.274110 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.274291 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.274426 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.274504 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:13Z","lastTransitionTime":"2025-12-04T10:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.377796 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.377856 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.377868 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.377891 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.377910 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:13Z","lastTransitionTime":"2025-12-04T10:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.479709 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.479763 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.479772 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.479790 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.479804 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:13Z","lastTransitionTime":"2025-12-04T10:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.564286 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.564326 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.564358 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.564362 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:13 crc kubenswrapper[4943]: E1204 10:17:13.564416 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:13 crc kubenswrapper[4943]: E1204 10:17:13.564555 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:13 crc kubenswrapper[4943]: E1204 10:17:13.564679 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:13 crc kubenswrapper[4943]: E1204 10:17:13.564806 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.581949 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.582012 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.582025 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.582044 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.582055 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:13Z","lastTransitionTime":"2025-12-04T10:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.685075 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.685111 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.685120 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.685135 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.685146 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:13Z","lastTransitionTime":"2025-12-04T10:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.787178 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.787237 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.787253 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.787269 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.787280 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:13Z","lastTransitionTime":"2025-12-04T10:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.890499 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.890547 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.890560 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.890575 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.890586 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:13Z","lastTransitionTime":"2025-12-04T10:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.992917 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.992956 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.992987 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.993009 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:13 crc kubenswrapper[4943]: I1204 10:17:13.993019 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:13Z","lastTransitionTime":"2025-12-04T10:17:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.095596 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.095651 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.095664 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.095681 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.095698 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:14Z","lastTransitionTime":"2025-12-04T10:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.198499 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.198545 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.198556 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.198575 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.198588 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:14Z","lastTransitionTime":"2025-12-04T10:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.300634 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.300686 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.300696 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.300710 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.300720 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:14Z","lastTransitionTime":"2025-12-04T10:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.402793 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.402840 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.402849 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.402865 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.402874 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:14Z","lastTransitionTime":"2025-12-04T10:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.504706 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.504744 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.504757 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.504771 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.504779 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:14Z","lastTransitionTime":"2025-12-04T10:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.606998 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.607264 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.607373 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.607448 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.607526 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:14Z","lastTransitionTime":"2025-12-04T10:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.709794 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.709841 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.709851 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.709868 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.709877 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:14Z","lastTransitionTime":"2025-12-04T10:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.813284 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.813330 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.813339 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.813355 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.813365 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:14Z","lastTransitionTime":"2025-12-04T10:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.917330 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.917387 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.917405 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.917429 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:14 crc kubenswrapper[4943]: I1204 10:17:14.917445 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:14Z","lastTransitionTime":"2025-12-04T10:17:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.020139 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.020295 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.020312 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.020337 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.020354 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:15Z","lastTransitionTime":"2025-12-04T10:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.123606 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.123671 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.123682 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.123698 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.123709 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:15Z","lastTransitionTime":"2025-12-04T10:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.226297 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.226335 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.226343 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.226360 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.226369 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:15Z","lastTransitionTime":"2025-12-04T10:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.330067 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.330226 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.330255 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.330280 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.330298 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:15Z","lastTransitionTime":"2025-12-04T10:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.363732 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.363790 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.363808 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.363832 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.363850 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:15Z","lastTransitionTime":"2025-12-04T10:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:15 crc kubenswrapper[4943]: E1204 10:17:15.377452 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.381534 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.381567 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.381576 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.381590 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.381630 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:15Z","lastTransitionTime":"2025-12-04T10:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:15 crc kubenswrapper[4943]: E1204 10:17:15.394873 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.398764 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.398798 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.398807 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.398820 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.398829 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:15Z","lastTransitionTime":"2025-12-04T10:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:15 crc kubenswrapper[4943]: E1204 10:17:15.411384 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.415058 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.415095 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.415106 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.415122 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.415133 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:15Z","lastTransitionTime":"2025-12-04T10:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:15 crc kubenswrapper[4943]: E1204 10:17:15.426287 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.429753 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.429812 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.429822 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.429838 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.429851 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:15Z","lastTransitionTime":"2025-12-04T10:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:15 crc kubenswrapper[4943]: E1204 10:17:15.443412 4943 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T10:17:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"caef8e39-80d0-412e-864d-57fe46732950\\\",\\\"systemUUID\\\":\\\"32a42f7c-963d-4042-9da3-416ccee17477\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:15Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:15 crc kubenswrapper[4943]: E1204 10:17:15.443574 4943 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.445368 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.445415 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.445431 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.445452 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.445465 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:15Z","lastTransitionTime":"2025-12-04T10:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.547528 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.547572 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.547580 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.547596 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.547605 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:15Z","lastTransitionTime":"2025-12-04T10:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.563924 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.563969 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.564010 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:15 crc kubenswrapper[4943]: E1204 10:17:15.564060 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.564083 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:15 crc kubenswrapper[4943]: E1204 10:17:15.564136 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:15 crc kubenswrapper[4943]: E1204 10:17:15.564169 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:15 crc kubenswrapper[4943]: E1204 10:17:15.564226 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.650097 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.650141 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.650151 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.650167 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:15 crc kubenswrapper[4943]: I1204 10:17:15.650178 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:15Z","lastTransitionTime":"2025-12-04T10:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.261369 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.261422 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.261432 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.261452 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.261464 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:16Z","lastTransitionTime":"2025-12-04T10:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.364009 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.364051 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.364062 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.364079 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.364093 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:16Z","lastTransitionTime":"2025-12-04T10:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.466522 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.466559 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.466567 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.466582 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.466590 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:16Z","lastTransitionTime":"2025-12-04T10:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.565245 4943 scope.go:117] "RemoveContainer" containerID="7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.568332 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.568409 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.568425 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.568441 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.568453 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:16Z","lastTransitionTime":"2025-12-04T10:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.578657 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.670613 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.670646 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.670658 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.670671 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.670680 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:16Z","lastTransitionTime":"2025-12-04T10:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.773808 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.773867 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.773884 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.773906 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.773924 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:16Z","lastTransitionTime":"2025-12-04T10:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.876832 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.876885 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.876896 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.876914 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.876926 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:16Z","lastTransitionTime":"2025-12-04T10:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.980599 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.980637 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.980655 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.980673 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:16 crc kubenswrapper[4943]: I1204 10:17:16.980690 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:16Z","lastTransitionTime":"2025-12-04T10:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.083260 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.083315 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.083325 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.083341 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.083351 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:17Z","lastTransitionTime":"2025-12-04T10:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.188749 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.188789 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.188798 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.188811 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.188819 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:17Z","lastTransitionTime":"2025-12-04T10:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.291258 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.291307 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.291323 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.291344 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.291355 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:17Z","lastTransitionTime":"2025-12-04T10:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.326481 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.326566 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.326607 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.326631 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.326743 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.326697022 +0000 UTC m=+149.915672910 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.326807 4943 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.326864 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.326896 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.326877356 +0000 UTC m=+149.915853224 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.326896 4943 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.327000 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.326971599 +0000 UTC m=+149.915947507 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.326907 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.327066 4943 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.327155 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.327133252 +0000 UTC m=+149.916109170 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.393950 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.394078 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.394104 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.394133 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.394154 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:17Z","lastTransitionTime":"2025-12-04T10:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.496290 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.496331 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.496343 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.496360 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.496371 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:17Z","lastTransitionTime":"2025-12-04T10:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.528988 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.529279 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.529298 4943 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.529309 4943 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.529355 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.529342069 +0000 UTC m=+150.118317937 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.564043 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.564133 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.564177 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.564359 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.564358 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.564474 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.564576 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:17 crc kubenswrapper[4943]: E1204 10:17:17.564641 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.599275 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.599319 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.599329 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.599345 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.599355 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:17Z","lastTransitionTime":"2025-12-04T10:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.701720 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.701757 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.701768 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.701790 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.701805 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:17Z","lastTransitionTime":"2025-12-04T10:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.804471 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.804502 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.804510 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.804523 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.804535 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:17Z","lastTransitionTime":"2025-12-04T10:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.906305 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.906573 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.906665 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.906768 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:17 crc kubenswrapper[4943]: I1204 10:17:17.906857 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:17Z","lastTransitionTime":"2025-12-04T10:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.015247 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.015317 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.015327 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.015343 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.015359 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:18Z","lastTransitionTime":"2025-12-04T10:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.119409 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.119455 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.119467 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.119490 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.119504 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:18Z","lastTransitionTime":"2025-12-04T10:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.222677 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.222709 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.222717 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.222733 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.222742 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:18Z","lastTransitionTime":"2025-12-04T10:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.326223 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.326268 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.326290 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.326322 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.326341 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:18Z","lastTransitionTime":"2025-12-04T10:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.428599 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.428659 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.428679 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.428696 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.428708 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:18Z","lastTransitionTime":"2025-12-04T10:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.463482 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/2.log" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.467030 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a"} Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.468114 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.497350 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5809e732-37d5-442a-8f7c-6cbaffe68de8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3db38524a3da9b4b5a7bf35d9329a0aa6a74d3bd3d7424feebb1cdabc8e043c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe322eaf4885153ea4ea0e0f059997517b5c4d83d102d4ccd426deedbded72e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48144828027b06c9f247dce0123e1b45080288632288017b26ce1906896822ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f18df637c05e1f517e5f5e41d9977e688f153198a7e144494c7ea3d31ed30f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://716ef8a0bf82dcbc707232cf9e7a72b09730d7cbd421950b792eb5e04face665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://141e6d35a5a43cc2dcd8b9c2c70d3745a9f53d92178b4ef6718dc776a6ee1a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://141e6d35a5a43cc2dcd8b9c2c70d3745a9f53d92178b4ef6718dc776a6ee1a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://18b4a76c7fed98c4f74e31ae34bd4948bea954e2a49ad118b35cbb5b26bc2faf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b4a76c7fed98c4f74e31ae34bd4948bea954e2a49ad118b35cbb5b26bc2faf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://9e881c9fc1b5a53b0c3e9e298071be7bdd6f674a3140cc4a6b1d0073b6ff2494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e881c9fc1b5a53b0c3e9e298071be7bdd6f674a3140cc4a6b1d0073b6ff2494\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.515118 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8c3544b-029d-48af-862b-a4c40406cfc7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:55Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T10:15:53Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.531243 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.531299 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.531310 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.531325 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.531347 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:18Z","lastTransitionTime":"2025-12-04T10:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.532502 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:14Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d3a1972d3f9fe278f4e59bb3aee3dc5bf67cf110f0211151a371e2d4593197be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.549116 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.562621 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:12Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.572993 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-lfcbx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5a911ae-5fad-4b9a-bfdd-5f2c7f1b33cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfdefc2f5ba2afb7ede0b4dd432ce39326c3e65b1cb0577f98600b120899c45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjxng\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-lfcbx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.584524 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5njp4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eeee576e-fae5-42f5-9248-0864812c01aa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d31a8e83f1275421097a3942d0b675a534ce69947c94a3a7d5198f1c4c7b73d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7p5w8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:20Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5njp4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.598348 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13e11ac-5073-4735-bc48-196c6ebeaeaa\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c73f41fccc9588fd128ebb062bf836fffcf812fbbb0106b609679eece6b9ddc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7600c4c6a4201d9f62b5375db68a7d5ecd8d5bafea3223956014e54d41a59f8a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dtxrq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v6hxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.608998 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f540258aa83206520a445cd14ff637e404d58967a896b4e47d613e8688cd9863\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:16:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qbwhk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:13Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-kswzd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.619795 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpblt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:16:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hw5pj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.633612 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.633659 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.633678 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.633697 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.633709 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:18Z","lastTransitionTime":"2025-12-04T10:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.634606 4943 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"faf53420-d3a8-47b2-b296-4ebbf6baa530\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:16:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T10:15:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0b2b59a76b99dc3fba5a9f52c9cdf1da4d8e1f8c98a6c6b325202c95dbd227e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52fc8f22eb736e7ad0680b7156d0a63d3973e51cdc9c9cba6a8ce94928a961e2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://490276a238d90dcae6fc74d20be3413c4d34d6325b8c006d7db9dc12a28b7465\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T10:15:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T10:15:52Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T10:17:18Z is after 2025-08-24T17:21:41Z" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.662553 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=34.662516423 podStartE2EDuration="34.662516423s" podCreationTimestamp="2025-12-04 10:16:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:18.6578176 +0000 UTC m=+87.246793478" watchObservedRunningTime="2025-12-04 10:17:18.662516423 +0000 UTC m=+87.251492291" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.697817 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-9nxvx" podStartSLOduration=65.697793058 podStartE2EDuration="1m5.697793058s" podCreationTimestamp="2025-12-04 10:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:18.697744457 +0000 UTC m=+87.286720315" watchObservedRunningTime="2025-12-04 10:17:18.697793058 +0000 UTC m=+87.286768926" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.736258 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.736528 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.736634 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.736715 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.736783 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:18Z","lastTransitionTime":"2025-12-04T10:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.744290 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-wkbf5" podStartSLOduration=65.744274993 podStartE2EDuration="1m5.744274993s" podCreationTimestamp="2025-12-04 10:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:18.743111006 +0000 UTC m=+87.332086874" watchObservedRunningTime="2025-12-04 10:17:18.744274993 +0000 UTC m=+87.333250861" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.848309 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.848341 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.848349 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.848363 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.848372 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:18Z","lastTransitionTime":"2025-12-04T10:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.950957 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.950983 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.950991 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.951004 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:18 crc kubenswrapper[4943]: I1204 10:17:18.951016 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:18Z","lastTransitionTime":"2025-12-04T10:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.054037 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.054097 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.054112 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.054129 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.054141 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:19Z","lastTransitionTime":"2025-12-04T10:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.156277 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.156319 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.156328 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.156346 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.156356 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:19Z","lastTransitionTime":"2025-12-04T10:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.258064 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.258104 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.258115 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.258131 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.258144 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:19Z","lastTransitionTime":"2025-12-04T10:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.360792 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.360843 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.360865 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.360882 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.360894 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:19Z","lastTransitionTime":"2025-12-04T10:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.463134 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.463178 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.463187 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.463220 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.463230 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:19Z","lastTransitionTime":"2025-12-04T10:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.474723 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/3.log" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.475371 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/2.log" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.477649 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a" exitCode=1 Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.477700 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a"} Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.477794 4943 scope.go:117] "RemoveContainer" containerID="7cc5604fa686f4671238a23c83cdc55966f1968fff3583e82779026496630a36" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.478474 4943 scope.go:117] "RemoveContainer" containerID="16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a" Dec 04 10:17:19 crc kubenswrapper[4943]: E1204 10:17:19.478734 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.516157 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v6hxd" podStartSLOduration=64.516136415 podStartE2EDuration="1m4.516136415s" podCreationTimestamp="2025-12-04 10:16:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:19.515645573 +0000 UTC m=+88.104621461" watchObservedRunningTime="2025-12-04 10:17:19.516136415 +0000 UTC m=+88.105112283" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.540188 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.540155091 podStartE2EDuration="3.540155091s" podCreationTimestamp="2025-12-04 10:17:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:19.53882786 +0000 UTC m=+88.127803738" watchObservedRunningTime="2025-12-04 10:17:19.540155091 +0000 UTC m=+88.129131029" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.566210 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.566182 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:19 crc kubenswrapper[4943]: E1204 10:17:19.566336 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.566351 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:19 crc kubenswrapper[4943]: E1204 10:17:19.566412 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:19 crc kubenswrapper[4943]: E1204 10:17:19.566487 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.566690 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:19 crc kubenswrapper[4943]: E1204 10:17:19.566863 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.567938 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.568050 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.568138 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.568239 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.568328 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:19Z","lastTransitionTime":"2025-12-04T10:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.573673 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=67.573651584 podStartE2EDuration="1m7.573651584s" podCreationTimestamp="2025-12-04 10:16:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:19.556759089 +0000 UTC m=+88.145734957" watchObservedRunningTime="2025-12-04 10:17:19.573651584 +0000 UTC m=+88.162627452" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.614079 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-lfcbx" podStartSLOduration=66.614056472 podStartE2EDuration="1m6.614056472s" podCreationTimestamp="2025-12-04 10:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:19.613792876 +0000 UTC m=+88.202768744" watchObservedRunningTime="2025-12-04 10:17:19.614056472 +0000 UTC m=+88.203032330" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.623698 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-5njp4" podStartSLOduration=65.623678543 podStartE2EDuration="1m5.623678543s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:19.623598262 +0000 UTC m=+88.212574140" watchObservedRunningTime="2025-12-04 10:17:19.623678543 +0000 UTC m=+88.212654411" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.648634 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podStartSLOduration=66.648613511 podStartE2EDuration="1m6.648613511s" podCreationTimestamp="2025-12-04 10:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:19.636106111 +0000 UTC m=+88.225081979" watchObservedRunningTime="2025-12-04 10:17:19.648613511 +0000 UTC m=+88.237589379" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.664142 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=59.664116092 podStartE2EDuration="59.664116092s" podCreationTimestamp="2025-12-04 10:16:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:19.664085732 +0000 UTC m=+88.253061610" watchObservedRunningTime="2025-12-04 10:17:19.664116092 +0000 UTC m=+88.253091960" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.671085 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.671120 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.671132 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.671145 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.671157 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:19Z","lastTransitionTime":"2025-12-04T10:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.773436 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.773479 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.773489 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.773528 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.773537 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:19Z","lastTransitionTime":"2025-12-04T10:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.875621 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.875673 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.875689 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.875707 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.875720 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:19Z","lastTransitionTime":"2025-12-04T10:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.978835 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.978897 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.978910 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.978923 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:19 crc kubenswrapper[4943]: I1204 10:17:19.978934 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:19Z","lastTransitionTime":"2025-12-04T10:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.081983 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.082094 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.082113 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.082136 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.082152 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:20Z","lastTransitionTime":"2025-12-04T10:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.185118 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.185155 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.185168 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.185186 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.185213 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:20Z","lastTransitionTime":"2025-12-04T10:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.287992 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.288284 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.288367 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.288489 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.288562 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:20Z","lastTransitionTime":"2025-12-04T10:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.391124 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.391195 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.391234 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.391261 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.391278 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:20Z","lastTransitionTime":"2025-12-04T10:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.482789 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/3.log" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.486987 4943 scope.go:117] "RemoveContainer" containerID="16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a" Dec 04 10:17:20 crc kubenswrapper[4943]: E1204 10:17:20.487169 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.498278 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.498307 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.498314 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.498328 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.498338 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:20Z","lastTransitionTime":"2025-12-04T10:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.601249 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.601292 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.601305 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.601322 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.601334 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:20Z","lastTransitionTime":"2025-12-04T10:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.703599 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.703645 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.703653 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.703671 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.703681 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:20Z","lastTransitionTime":"2025-12-04T10:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.806104 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.806157 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.806170 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.806190 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.806239 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:20Z","lastTransitionTime":"2025-12-04T10:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.909817 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.909879 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.909891 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.909907 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:20 crc kubenswrapper[4943]: I1204 10:17:20.909920 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:20Z","lastTransitionTime":"2025-12-04T10:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.012215 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.012247 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.012258 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.012273 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.012282 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:21Z","lastTransitionTime":"2025-12-04T10:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.114403 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.114439 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.114448 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.114461 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.114469 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:21Z","lastTransitionTime":"2025-12-04T10:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.216707 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.217039 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.217048 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.217062 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.217072 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:21Z","lastTransitionTime":"2025-12-04T10:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.319376 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.319428 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.319441 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.319463 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.319527 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:21Z","lastTransitionTime":"2025-12-04T10:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.422445 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.422484 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.422493 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.422506 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.422515 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:21Z","lastTransitionTime":"2025-12-04T10:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.524929 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.524975 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.524986 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.525028 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.525045 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:21Z","lastTransitionTime":"2025-12-04T10:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.564380 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.564417 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.564507 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.564737 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:21 crc kubenswrapper[4943]: E1204 10:17:21.564727 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:21 crc kubenswrapper[4943]: E1204 10:17:21.564856 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:21 crc kubenswrapper[4943]: E1204 10:17:21.564898 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:21 crc kubenswrapper[4943]: E1204 10:17:21.564969 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.627160 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.627224 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.627236 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.627253 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.627267 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:21Z","lastTransitionTime":"2025-12-04T10:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.730391 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.730446 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.730457 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.730473 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.730485 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:21Z","lastTransitionTime":"2025-12-04T10:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.832628 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.832664 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.832674 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.832692 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.832704 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:21Z","lastTransitionTime":"2025-12-04T10:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.935286 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.935327 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.935336 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.935350 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:21 crc kubenswrapper[4943]: I1204 10:17:21.935358 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:21Z","lastTransitionTime":"2025-12-04T10:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.037367 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.037435 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.037456 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.037487 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.037508 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:22Z","lastTransitionTime":"2025-12-04T10:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.140131 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.140190 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.140247 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.140273 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.140290 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:22Z","lastTransitionTime":"2025-12-04T10:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.242928 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.242973 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.242985 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.243001 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.243011 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:22Z","lastTransitionTime":"2025-12-04T10:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.345562 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.345606 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.345619 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.345637 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.345649 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:22Z","lastTransitionTime":"2025-12-04T10:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.448303 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.448346 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.448355 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.448369 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.448379 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:22Z","lastTransitionTime":"2025-12-04T10:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.550455 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.550492 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.550504 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.550526 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.550538 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:22Z","lastTransitionTime":"2025-12-04T10:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.654152 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.654235 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.654309 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.654333 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.654346 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:22Z","lastTransitionTime":"2025-12-04T10:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.756964 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.757007 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.757022 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.757100 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.757113 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:22Z","lastTransitionTime":"2025-12-04T10:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.859800 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.859835 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.859845 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.859863 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.859874 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:22Z","lastTransitionTime":"2025-12-04T10:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.962494 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.962552 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.962561 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.962574 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:22 crc kubenswrapper[4943]: I1204 10:17:22.962585 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:22Z","lastTransitionTime":"2025-12-04T10:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.064624 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.064653 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.064663 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.064677 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.064686 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:23Z","lastTransitionTime":"2025-12-04T10:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.167426 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.167464 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.167472 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.167488 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.167499 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:23Z","lastTransitionTime":"2025-12-04T10:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.270616 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.270683 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.270697 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.270716 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.270732 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:23Z","lastTransitionTime":"2025-12-04T10:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.373119 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.373176 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.373189 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.373282 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.373298 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:23Z","lastTransitionTime":"2025-12-04T10:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.475620 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.475738 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.475760 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.475784 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.475800 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:23Z","lastTransitionTime":"2025-12-04T10:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.564550 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.564606 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.564606 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:23 crc kubenswrapper[4943]: E1204 10:17:23.564755 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.564858 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:23 crc kubenswrapper[4943]: E1204 10:17:23.564914 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:23 crc kubenswrapper[4943]: E1204 10:17:23.565057 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:23 crc kubenswrapper[4943]: E1204 10:17:23.565174 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.579266 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.579333 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.579346 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.579370 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.579388 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:23Z","lastTransitionTime":"2025-12-04T10:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.681917 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.681958 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.681988 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.682002 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.682012 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:23Z","lastTransitionTime":"2025-12-04T10:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.784810 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.784965 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.785041 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.785069 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.785087 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:23Z","lastTransitionTime":"2025-12-04T10:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.887605 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.887639 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.887647 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.887663 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.887674 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:23Z","lastTransitionTime":"2025-12-04T10:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.990495 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.990535 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.990547 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.990564 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:23 crc kubenswrapper[4943]: I1204 10:17:23.990576 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:23Z","lastTransitionTime":"2025-12-04T10:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.092919 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.092973 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.092985 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.093001 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.093013 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:24Z","lastTransitionTime":"2025-12-04T10:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.195587 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.195640 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.195653 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.195669 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.195683 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:24Z","lastTransitionTime":"2025-12-04T10:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.298702 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.298768 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.298783 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.298842 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.298855 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:24Z","lastTransitionTime":"2025-12-04T10:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.401691 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.401729 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.401739 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.401754 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.401766 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:24Z","lastTransitionTime":"2025-12-04T10:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.503578 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.503616 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.503627 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.503642 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.503653 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:24Z","lastTransitionTime":"2025-12-04T10:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.606273 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.606321 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.606334 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.606353 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.606364 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:24Z","lastTransitionTime":"2025-12-04T10:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.708513 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.708552 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.708564 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.708579 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.708588 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:24Z","lastTransitionTime":"2025-12-04T10:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.810778 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.810844 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.810862 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.810885 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.810904 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:24Z","lastTransitionTime":"2025-12-04T10:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.913504 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.913553 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.913562 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.913583 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:24 crc kubenswrapper[4943]: I1204 10:17:24.913593 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:24Z","lastTransitionTime":"2025-12-04T10:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.016498 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.016550 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.016561 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.016581 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.016592 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:25Z","lastTransitionTime":"2025-12-04T10:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.118761 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.118803 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.118814 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.118830 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.118843 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:25Z","lastTransitionTime":"2025-12-04T10:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.223079 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.223137 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.223148 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.223166 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.223177 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:25Z","lastTransitionTime":"2025-12-04T10:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.324928 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.324966 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.324975 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.324988 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.324996 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:25Z","lastTransitionTime":"2025-12-04T10:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.427945 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.427990 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.427999 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.428015 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.428025 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:25Z","lastTransitionTime":"2025-12-04T10:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.529972 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.530038 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.530050 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.530069 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.530080 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:25Z","lastTransitionTime":"2025-12-04T10:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.563999 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.564101 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.564136 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:25 crc kubenswrapper[4943]: E1204 10:17:25.564187 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.564024 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:25 crc kubenswrapper[4943]: E1204 10:17:25.564395 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:25 crc kubenswrapper[4943]: E1204 10:17:25.564514 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:25 crc kubenswrapper[4943]: E1204 10:17:25.564600 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.633851 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.633907 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.633922 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.633942 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.633952 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:25Z","lastTransitionTime":"2025-12-04T10:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.737230 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.737291 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.737301 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.737318 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.737328 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:25Z","lastTransitionTime":"2025-12-04T10:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.766995 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.767042 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.767051 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.767068 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.767079 4943 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T10:17:25Z","lastTransitionTime":"2025-12-04T10:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.815149 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8"] Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.815808 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.818906 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.818941 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.821150 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.821179 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.922333 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9e570a2c-e30d-455a-a93c-5f250190c31e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.922400 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e570a2c-e30d-455a-a93c-5f250190c31e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.922432 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e570a2c-e30d-455a-a93c-5f250190c31e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.922469 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9e570a2c-e30d-455a-a93c-5f250190c31e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:25 crc kubenswrapper[4943]: I1204 10:17:25.922495 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9e570a2c-e30d-455a-a93c-5f250190c31e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.023656 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9e570a2c-e30d-455a-a93c-5f250190c31e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.023710 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e570a2c-e30d-455a-a93c-5f250190c31e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.023743 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e570a2c-e30d-455a-a93c-5f250190c31e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.023762 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9e570a2c-e30d-455a-a93c-5f250190c31e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.023780 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9e570a2c-e30d-455a-a93c-5f250190c31e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.023828 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9e570a2c-e30d-455a-a93c-5f250190c31e-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.023883 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9e570a2c-e30d-455a-a93c-5f250190c31e-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.024742 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9e570a2c-e30d-455a-a93c-5f250190c31e-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.029190 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e570a2c-e30d-455a-a93c-5f250190c31e-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.040053 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e570a2c-e30d-455a-a93c-5f250190c31e-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mr8d8\" (UID: \"9e570a2c-e30d-455a-a93c-5f250190c31e\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.131176 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" Dec 04 10:17:26 crc kubenswrapper[4943]: W1204 10:17:26.144362 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e570a2c_e30d_455a_a93c_5f250190c31e.slice/crio-de64160b71b1637a5c05dc19fcfdf08864bdad541b00c169ceed56d5d2e4ced8 WatchSource:0}: Error finding container de64160b71b1637a5c05dc19fcfdf08864bdad541b00c169ceed56d5d2e4ced8: Status 404 returned error can't find the container with id de64160b71b1637a5c05dc19fcfdf08864bdad541b00c169ceed56d5d2e4ced8 Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.507865 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" event={"ID":"9e570a2c-e30d-455a-a93c-5f250190c31e","Type":"ContainerStarted","Data":"61e278b3f088299aa76c74af209e3a37ba80ea3d01c8e839b130f324d8102476"} Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.507928 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" event={"ID":"9e570a2c-e30d-455a-a93c-5f250190c31e","Type":"ContainerStarted","Data":"de64160b71b1637a5c05dc19fcfdf08864bdad541b00c169ceed56d5d2e4ced8"} Dec 04 10:17:26 crc kubenswrapper[4943]: I1204 10:17:26.527897 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mr8d8" podStartSLOduration=73.527875344 podStartE2EDuration="1m13.527875344s" podCreationTimestamp="2025-12-04 10:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:26.526217895 +0000 UTC m=+95.115193773" watchObservedRunningTime="2025-12-04 10:17:26.527875344 +0000 UTC m=+95.116851232" Dec 04 10:17:27 crc kubenswrapper[4943]: I1204 10:17:27.564129 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:27 crc kubenswrapper[4943]: I1204 10:17:27.564129 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:27 crc kubenswrapper[4943]: I1204 10:17:27.564329 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:27 crc kubenswrapper[4943]: E1204 10:17:27.564500 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:27 crc kubenswrapper[4943]: E1204 10:17:27.564641 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:27 crc kubenswrapper[4943]: E1204 10:17:27.564891 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:27 crc kubenswrapper[4943]: I1204 10:17:27.564326 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:27 crc kubenswrapper[4943]: E1204 10:17:27.565027 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:27 crc kubenswrapper[4943]: I1204 10:17:27.576402 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 04 10:17:29 crc kubenswrapper[4943]: I1204 10:17:29.563719 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:29 crc kubenswrapper[4943]: E1204 10:17:29.563878 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:29 crc kubenswrapper[4943]: I1204 10:17:29.564122 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:29 crc kubenswrapper[4943]: E1204 10:17:29.564258 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:29 crc kubenswrapper[4943]: I1204 10:17:29.564429 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:29 crc kubenswrapper[4943]: I1204 10:17:29.564462 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:29 crc kubenswrapper[4943]: E1204 10:17:29.564506 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:29 crc kubenswrapper[4943]: E1204 10:17:29.564601 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:31 crc kubenswrapper[4943]: I1204 10:17:31.563939 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:31 crc kubenswrapper[4943]: I1204 10:17:31.563960 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:31 crc kubenswrapper[4943]: E1204 10:17:31.564475 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:31 crc kubenswrapper[4943]: I1204 10:17:31.563988 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:31 crc kubenswrapper[4943]: E1204 10:17:31.564565 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:31 crc kubenswrapper[4943]: I1204 10:17:31.563961 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:31 crc kubenswrapper[4943]: E1204 10:17:31.564649 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:31 crc kubenswrapper[4943]: E1204 10:17:31.564716 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:31 crc kubenswrapper[4943]: I1204 10:17:31.564830 4943 scope.go:117] "RemoveContainer" containerID="16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a" Dec 04 10:17:31 crc kubenswrapper[4943]: E1204 10:17:31.564989 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" Dec 04 10:17:32 crc kubenswrapper[4943]: I1204 10:17:32.289304 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:32 crc kubenswrapper[4943]: E1204 10:17:32.289593 4943 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:17:32 crc kubenswrapper[4943]: E1204 10:17:32.289782 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs podName:eb85a8e0-6c90-47cb-a1b5-9ecd5244c710 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:36.289723452 +0000 UTC m=+164.878699400 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs") pod "network-metrics-daemon-hw5pj" (UID: "eb85a8e0-6c90-47cb-a1b5-9ecd5244c710") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 10:17:32 crc kubenswrapper[4943]: I1204 10:17:32.577021 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=5.576999688 podStartE2EDuration="5.576999688s" podCreationTimestamp="2025-12-04 10:17:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:32.576588959 +0000 UTC m=+101.165564827" watchObservedRunningTime="2025-12-04 10:17:32.576999688 +0000 UTC m=+101.165975576" Dec 04 10:17:33 crc kubenswrapper[4943]: I1204 10:17:33.564317 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:33 crc kubenswrapper[4943]: E1204 10:17:33.564459 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:33 crc kubenswrapper[4943]: I1204 10:17:33.564672 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:33 crc kubenswrapper[4943]: E1204 10:17:33.564741 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:33 crc kubenswrapper[4943]: I1204 10:17:33.564874 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:33 crc kubenswrapper[4943]: E1204 10:17:33.564971 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:33 crc kubenswrapper[4943]: I1204 10:17:33.565181 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:33 crc kubenswrapper[4943]: E1204 10:17:33.565309 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:35 crc kubenswrapper[4943]: I1204 10:17:35.563983 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:35 crc kubenswrapper[4943]: I1204 10:17:35.564023 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:35 crc kubenswrapper[4943]: I1204 10:17:35.563991 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:35 crc kubenswrapper[4943]: E1204 10:17:35.564287 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:35 crc kubenswrapper[4943]: E1204 10:17:35.564165 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:35 crc kubenswrapper[4943]: E1204 10:17:35.564481 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:35 crc kubenswrapper[4943]: I1204 10:17:35.565462 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:35 crc kubenswrapper[4943]: E1204 10:17:35.565866 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:37 crc kubenswrapper[4943]: I1204 10:17:37.564569 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:37 crc kubenswrapper[4943]: E1204 10:17:37.564794 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:37 crc kubenswrapper[4943]: I1204 10:17:37.564893 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:37 crc kubenswrapper[4943]: I1204 10:17:37.564970 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:37 crc kubenswrapper[4943]: I1204 10:17:37.564972 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:37 crc kubenswrapper[4943]: E1204 10:17:37.565134 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:37 crc kubenswrapper[4943]: E1204 10:17:37.565277 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:37 crc kubenswrapper[4943]: E1204 10:17:37.565435 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:39 crc kubenswrapper[4943]: I1204 10:17:39.564801 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:39 crc kubenswrapper[4943]: I1204 10:17:39.564802 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:39 crc kubenswrapper[4943]: I1204 10:17:39.564984 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:39 crc kubenswrapper[4943]: I1204 10:17:39.564822 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:39 crc kubenswrapper[4943]: E1204 10:17:39.565305 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:39 crc kubenswrapper[4943]: E1204 10:17:39.565973 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:39 crc kubenswrapper[4943]: E1204 10:17:39.566086 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:39 crc kubenswrapper[4943]: E1204 10:17:39.566156 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:41 crc kubenswrapper[4943]: I1204 10:17:41.563654 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:41 crc kubenswrapper[4943]: E1204 10:17:41.563840 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:41 crc kubenswrapper[4943]: I1204 10:17:41.564320 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:41 crc kubenswrapper[4943]: E1204 10:17:41.564422 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:41 crc kubenswrapper[4943]: I1204 10:17:41.570496 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:41 crc kubenswrapper[4943]: I1204 10:17:41.570507 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:41 crc kubenswrapper[4943]: E1204 10:17:41.570716 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:41 crc kubenswrapper[4943]: E1204 10:17:41.570871 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:43 crc kubenswrapper[4943]: I1204 10:17:43.564268 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:43 crc kubenswrapper[4943]: I1204 10:17:43.564402 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:43 crc kubenswrapper[4943]: I1204 10:17:43.564402 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:43 crc kubenswrapper[4943]: I1204 10:17:43.564719 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:43 crc kubenswrapper[4943]: E1204 10:17:43.564789 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:43 crc kubenswrapper[4943]: E1204 10:17:43.564888 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:43 crc kubenswrapper[4943]: E1204 10:17:43.564604 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:43 crc kubenswrapper[4943]: E1204 10:17:43.564963 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:45 crc kubenswrapper[4943]: I1204 10:17:45.564262 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:45 crc kubenswrapper[4943]: E1204 10:17:45.564488 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:45 crc kubenswrapper[4943]: I1204 10:17:45.564697 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:45 crc kubenswrapper[4943]: E1204 10:17:45.564763 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:45 crc kubenswrapper[4943]: I1204 10:17:45.564886 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:45 crc kubenswrapper[4943]: E1204 10:17:45.564943 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:45 crc kubenswrapper[4943]: I1204 10:17:45.565049 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:45 crc kubenswrapper[4943]: E1204 10:17:45.565100 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:46 crc kubenswrapper[4943]: I1204 10:17:46.566337 4943 scope.go:117] "RemoveContainer" containerID="16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a" Dec 04 10:17:46 crc kubenswrapper[4943]: E1204 10:17:46.566706 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-7g4wf_openshift-ovn-kubernetes(d0498f58-c99b-4fa8-a404-f7f43199c29f)\"" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" Dec 04 10:17:47 crc kubenswrapper[4943]: I1204 10:17:47.564027 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:47 crc kubenswrapper[4943]: I1204 10:17:47.564072 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:47 crc kubenswrapper[4943]: I1204 10:17:47.564094 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:47 crc kubenswrapper[4943]: E1204 10:17:47.564195 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:47 crc kubenswrapper[4943]: E1204 10:17:47.564374 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:47 crc kubenswrapper[4943]: E1204 10:17:47.564541 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:47 crc kubenswrapper[4943]: I1204 10:17:47.564634 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:47 crc kubenswrapper[4943]: E1204 10:17:47.564806 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:49 crc kubenswrapper[4943]: I1204 10:17:49.564625 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:49 crc kubenswrapper[4943]: I1204 10:17:49.564709 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:49 crc kubenswrapper[4943]: E1204 10:17:49.564818 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:49 crc kubenswrapper[4943]: I1204 10:17:49.564838 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:49 crc kubenswrapper[4943]: I1204 10:17:49.564645 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:49 crc kubenswrapper[4943]: E1204 10:17:49.565008 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:49 crc kubenswrapper[4943]: E1204 10:17:49.565150 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:49 crc kubenswrapper[4943]: E1204 10:17:49.565337 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:49 crc kubenswrapper[4943]: I1204 10:17:49.598674 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wkbf5_b9cb494f-7d4c-45f2-8b9b-e35c42b41c79/kube-multus/1.log" Dec 04 10:17:49 crc kubenswrapper[4943]: I1204 10:17:49.599296 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wkbf5_b9cb494f-7d4c-45f2-8b9b-e35c42b41c79/kube-multus/0.log" Dec 04 10:17:49 crc kubenswrapper[4943]: I1204 10:17:49.599389 4943 generic.go:334] "Generic (PLEG): container finished" podID="b9cb494f-7d4c-45f2-8b9b-e35c42b41c79" containerID="635d7d730273976acb953de2147ab4353647915f5f9898a96c04558ab8262a1d" exitCode=1 Dec 04 10:17:49 crc kubenswrapper[4943]: I1204 10:17:49.599451 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wkbf5" event={"ID":"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79","Type":"ContainerDied","Data":"635d7d730273976acb953de2147ab4353647915f5f9898a96c04558ab8262a1d"} Dec 04 10:17:49 crc kubenswrapper[4943]: I1204 10:17:49.599529 4943 scope.go:117] "RemoveContainer" containerID="babe766250ae05116a994339f7cbac6441f8ccfb1491877ec4da05274fcae793" Dec 04 10:17:49 crc kubenswrapper[4943]: I1204 10:17:49.600311 4943 scope.go:117] "RemoveContainer" containerID="635d7d730273976acb953de2147ab4353647915f5f9898a96c04558ab8262a1d" Dec 04 10:17:49 crc kubenswrapper[4943]: E1204 10:17:49.600684 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-wkbf5_openshift-multus(b9cb494f-7d4c-45f2-8b9b-e35c42b41c79)\"" pod="openshift-multus/multus-wkbf5" podUID="b9cb494f-7d4c-45f2-8b9b-e35c42b41c79" Dec 04 10:17:50 crc kubenswrapper[4943]: I1204 10:17:50.605556 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wkbf5_b9cb494f-7d4c-45f2-8b9b-e35c42b41c79/kube-multus/1.log" Dec 04 10:17:51 crc kubenswrapper[4943]: I1204 10:17:51.564378 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:51 crc kubenswrapper[4943]: I1204 10:17:51.564413 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:51 crc kubenswrapper[4943]: E1204 10:17:51.564577 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:51 crc kubenswrapper[4943]: I1204 10:17:51.564665 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:51 crc kubenswrapper[4943]: E1204 10:17:51.564810 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:51 crc kubenswrapper[4943]: E1204 10:17:51.565556 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:51 crc kubenswrapper[4943]: I1204 10:17:51.565699 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:51 crc kubenswrapper[4943]: E1204 10:17:51.565835 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:52 crc kubenswrapper[4943]: E1204 10:17:52.080560 4943 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 04 10:17:52 crc kubenswrapper[4943]: E1204 10:17:52.685790 4943 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 10:17:53 crc kubenswrapper[4943]: I1204 10:17:53.564260 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:53 crc kubenswrapper[4943]: I1204 10:17:53.564282 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:53 crc kubenswrapper[4943]: E1204 10:17:53.564406 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:53 crc kubenswrapper[4943]: I1204 10:17:53.564290 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:53 crc kubenswrapper[4943]: E1204 10:17:53.564538 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:53 crc kubenswrapper[4943]: I1204 10:17:53.564289 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:53 crc kubenswrapper[4943]: E1204 10:17:53.564704 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:53 crc kubenswrapper[4943]: E1204 10:17:53.564638 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:55 crc kubenswrapper[4943]: I1204 10:17:55.564403 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:55 crc kubenswrapper[4943]: I1204 10:17:55.564468 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:55 crc kubenswrapper[4943]: I1204 10:17:55.564525 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:55 crc kubenswrapper[4943]: I1204 10:17:55.564428 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:55 crc kubenswrapper[4943]: E1204 10:17:55.564566 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:55 crc kubenswrapper[4943]: E1204 10:17:55.564647 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:55 crc kubenswrapper[4943]: E1204 10:17:55.564753 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:55 crc kubenswrapper[4943]: E1204 10:17:55.564869 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:57 crc kubenswrapper[4943]: I1204 10:17:57.564494 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:57 crc kubenswrapper[4943]: I1204 10:17:57.564609 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:57 crc kubenswrapper[4943]: E1204 10:17:57.564641 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:57 crc kubenswrapper[4943]: I1204 10:17:57.564666 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:57 crc kubenswrapper[4943]: I1204 10:17:57.564816 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:57 crc kubenswrapper[4943]: E1204 10:17:57.564906 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:57 crc kubenswrapper[4943]: E1204 10:17:57.564936 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:57 crc kubenswrapper[4943]: E1204 10:17:57.564986 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:57 crc kubenswrapper[4943]: E1204 10:17:57.687430 4943 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 10:17:59 crc kubenswrapper[4943]: I1204 10:17:59.564529 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:17:59 crc kubenswrapper[4943]: I1204 10:17:59.564561 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:17:59 crc kubenswrapper[4943]: I1204 10:17:59.564699 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:17:59 crc kubenswrapper[4943]: E1204 10:17:59.564971 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:17:59 crc kubenswrapper[4943]: I1204 10:17:59.565309 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:17:59 crc kubenswrapper[4943]: E1204 10:17:59.565537 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:17:59 crc kubenswrapper[4943]: E1204 10:17:59.565884 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:17:59 crc kubenswrapper[4943]: E1204 10:17:59.565706 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:17:59 crc kubenswrapper[4943]: I1204 10:17:59.566659 4943 scope.go:117] "RemoveContainer" containerID="16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a" Dec 04 10:18:00 crc kubenswrapper[4943]: I1204 10:18:00.646485 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/3.log" Dec 04 10:18:00 crc kubenswrapper[4943]: I1204 10:18:00.649823 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerStarted","Data":"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce"} Dec 04 10:18:00 crc kubenswrapper[4943]: I1204 10:18:00.650330 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:18:00 crc kubenswrapper[4943]: I1204 10:18:00.699416 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podStartSLOduration=107.699374245 podStartE2EDuration="1m47.699374245s" podCreationTimestamp="2025-12-04 10:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:17:18.767625502 +0000 UTC m=+87.356601380" watchObservedRunningTime="2025-12-04 10:18:00.699374245 +0000 UTC m=+129.288350113" Dec 04 10:18:01 crc kubenswrapper[4943]: I1204 10:18:01.299161 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hw5pj"] Dec 04 10:18:01 crc kubenswrapper[4943]: I1204 10:18:01.299399 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:18:01 crc kubenswrapper[4943]: E1204 10:18:01.299559 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:18:01 crc kubenswrapper[4943]: I1204 10:18:01.564450 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:18:01 crc kubenswrapper[4943]: I1204 10:18:01.564515 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:18:01 crc kubenswrapper[4943]: I1204 10:18:01.564566 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:18:01 crc kubenswrapper[4943]: E1204 10:18:01.564594 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:18:01 crc kubenswrapper[4943]: E1204 10:18:01.564754 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:18:01 crc kubenswrapper[4943]: E1204 10:18:01.564797 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:18:01 crc kubenswrapper[4943]: I1204 10:18:01.564906 4943 scope.go:117] "RemoveContainer" containerID="635d7d730273976acb953de2147ab4353647915f5f9898a96c04558ab8262a1d" Dec 04 10:18:02 crc kubenswrapper[4943]: I1204 10:18:02.565703 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:18:02 crc kubenswrapper[4943]: E1204 10:18:02.565804 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:18:02 crc kubenswrapper[4943]: I1204 10:18:02.657790 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wkbf5_b9cb494f-7d4c-45f2-8b9b-e35c42b41c79/kube-multus/1.log" Dec 04 10:18:02 crc kubenswrapper[4943]: I1204 10:18:02.657835 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wkbf5" event={"ID":"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79","Type":"ContainerStarted","Data":"f48e036e9d741d48766b670629348a5e5592c0f1ef59ed8f5059e0df75057a5a"} Dec 04 10:18:02 crc kubenswrapper[4943]: E1204 10:18:02.688421 4943 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 10:18:03 crc kubenswrapper[4943]: I1204 10:18:03.564329 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:18:03 crc kubenswrapper[4943]: I1204 10:18:03.564349 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:18:03 crc kubenswrapper[4943]: I1204 10:18:03.564367 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:18:03 crc kubenswrapper[4943]: E1204 10:18:03.565864 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:18:03 crc kubenswrapper[4943]: E1204 10:18:03.565980 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:18:03 crc kubenswrapper[4943]: E1204 10:18:03.566088 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:18:04 crc kubenswrapper[4943]: I1204 10:18:04.564488 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:18:04 crc kubenswrapper[4943]: E1204 10:18:04.564985 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:18:05 crc kubenswrapper[4943]: I1204 10:18:05.564424 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:18:05 crc kubenswrapper[4943]: E1204 10:18:05.565102 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:18:05 crc kubenswrapper[4943]: I1204 10:18:05.564440 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:18:05 crc kubenswrapper[4943]: E1204 10:18:05.565185 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:18:05 crc kubenswrapper[4943]: I1204 10:18:05.564424 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:18:05 crc kubenswrapper[4943]: E1204 10:18:05.565267 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:18:06 crc kubenswrapper[4943]: I1204 10:18:06.563882 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:18:06 crc kubenswrapper[4943]: E1204 10:18:06.564125 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hw5pj" podUID="eb85a8e0-6c90-47cb-a1b5-9ecd5244c710" Dec 04 10:18:07 crc kubenswrapper[4943]: I1204 10:18:07.564555 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:18:07 crc kubenswrapper[4943]: I1204 10:18:07.564674 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:18:07 crc kubenswrapper[4943]: E1204 10:18:07.564836 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 10:18:07 crc kubenswrapper[4943]: I1204 10:18:07.565103 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:18:07 crc kubenswrapper[4943]: E1204 10:18:07.565231 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 10:18:07 crc kubenswrapper[4943]: E1204 10:18:07.565475 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 10:18:08 crc kubenswrapper[4943]: I1204 10:18:08.564728 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:18:08 crc kubenswrapper[4943]: I1204 10:18:08.567722 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 10:18:08 crc kubenswrapper[4943]: I1204 10:18:08.568356 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 10:18:09 crc kubenswrapper[4943]: I1204 10:18:09.564560 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:18:09 crc kubenswrapper[4943]: I1204 10:18:09.565194 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:18:09 crc kubenswrapper[4943]: I1204 10:18:09.565571 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:18:09 crc kubenswrapper[4943]: I1204 10:18:09.571539 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 10:18:09 crc kubenswrapper[4943]: I1204 10:18:09.571691 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 10:18:09 crc kubenswrapper[4943]: I1204 10:18:09.571883 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 10:18:09 crc kubenswrapper[4943]: I1204 10:18:09.572158 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 10:18:16 crc kubenswrapper[4943]: I1204 10:18:16.956847 4943 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.003311 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-8sdsb"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.004431 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2g96c"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.004561 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.004850 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.008110 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.008290 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.008482 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.017249 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.018012 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.018968 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.023587 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.023693 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.023780 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.024218 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.024680 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.031893 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.032504 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.032633 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.037609 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.037746 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.037939 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.038130 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.038188 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.038295 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.038421 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.038568 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.038676 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.038753 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.045996 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.046220 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.047061 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.047401 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.047620 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.048962 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.049276 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-grw5p"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.049896 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-grw5p" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.050812 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.057954 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.060325 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.062488 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.062662 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.062764 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.062950 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.063170 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.063283 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.063374 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.063404 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064304 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b66208b0-ad95-479b-8fae-c239c7d91fdb-audit-policies\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064335 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b66208b0-ad95-479b-8fae-c239c7d91fdb-audit-dir\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064381 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58b0ae05-55dc-4238-9163-c8e06974ffca-config\") pod \"route-controller-manager-6576b87f9c-h9dk9\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064408 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f2720fe-737e-4562-9fc1-d0e52e0d99a2-config\") pod \"machine-api-operator-5694c8668f-8sdsb\" (UID: \"1f2720fe-737e-4562-9fc1-d0e52e0d99a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064422 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b66208b0-ad95-479b-8fae-c239c7d91fdb-serving-cert\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064446 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b66208b0-ad95-479b-8fae-c239c7d91fdb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064465 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1f2720fe-737e-4562-9fc1-d0e52e0d99a2-images\") pod \"machine-api-operator-5694c8668f-8sdsb\" (UID: \"1f2720fe-737e-4562-9fc1-d0e52e0d99a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064507 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d85wc\" (UniqueName: \"kubernetes.io/projected/1f2720fe-737e-4562-9fc1-d0e52e0d99a2-kube-api-access-d85wc\") pod \"machine-api-operator-5694c8668f-8sdsb\" (UID: \"1f2720fe-737e-4562-9fc1-d0e52e0d99a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064528 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8490c962-38c0-4be5-b940-d4c4d08c8a55-serving-cert\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064547 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j72zg\" (UniqueName: \"kubernetes.io/projected/8490c962-38c0-4be5-b940-d4c4d08c8a55-kube-api-access-j72zg\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064563 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6ssz\" (UniqueName: \"kubernetes.io/projected/58b0ae05-55dc-4238-9163-c8e06974ffca-kube-api-access-w6ssz\") pod \"route-controller-manager-6576b87f9c-h9dk9\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064568 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-28rrg"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064888 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.064579 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-client-ca\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065385 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b66208b0-ad95-479b-8fae-c239c7d91fdb-etcd-client\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065402 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k99j\" (UniqueName: \"kubernetes.io/projected/b66208b0-ad95-479b-8fae-c239c7d91fdb-kube-api-access-6k99j\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065484 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfll6\" (UniqueName: \"kubernetes.io/projected/fd459946-dd06-42f3-896c-8ae5b7394b33-kube-api-access-rfll6\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbf9m\" (UID: \"fd459946-dd06-42f3-896c-8ae5b7394b33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065503 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58b0ae05-55dc-4238-9163-c8e06974ffca-serving-cert\") pod \"route-controller-manager-6576b87f9c-h9dk9\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065519 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd459946-dd06-42f3-896c-8ae5b7394b33-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbf9m\" (UID: \"fd459946-dd06-42f3-896c-8ae5b7394b33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065536 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b66208b0-ad95-479b-8fae-c239c7d91fdb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065551 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065566 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd459946-dd06-42f3-896c-8ae5b7394b33-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbf9m\" (UID: \"fd459946-dd06-42f3-896c-8ae5b7394b33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065685 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b66208b0-ad95-479b-8fae-c239c7d91fdb-encryption-config\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065722 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-config\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065762 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58b0ae05-55dc-4238-9163-c8e06974ffca-client-ca\") pod \"route-controller-manager-6576b87f9c-h9dk9\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065812 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.065782 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1f2720fe-737e-4562-9fc1-d0e52e0d99a2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8sdsb\" (UID: \"1f2720fe-737e-4562-9fc1-d0e52e0d99a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.066292 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wkldv"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.066564 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.067015 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.067994 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.068354 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.068455 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-28rrg" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.072103 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.072321 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.072864 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.073068 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.073304 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.079728 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.080436 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.083665 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.083938 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.084077 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.084367 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.084658 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.084774 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.084880 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.084980 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.085232 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.085436 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.086082 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5bgm4"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.087129 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-kdt7l"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.087573 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.087634 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.087855 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.088292 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.089021 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ml4z6"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.089713 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ml4z6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.091116 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.113530 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.117529 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hgd88"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.119520 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.120253 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.129751 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-tf5g6"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.130639 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.134858 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.138490 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.138786 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.138854 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.138994 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.139058 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.139193 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.139919 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.140274 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.140435 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.140945 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.141391 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.141472 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.141636 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.141849 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.141894 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.142033 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.142116 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.142253 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.142409 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.142512 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.142542 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.142687 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.142787 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.142814 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.142981 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.143224 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.143440 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.143672 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.143722 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.143735 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.143862 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.143902 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.143964 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.144084 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.144178 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.144350 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.144376 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.144463 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.144634 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.144653 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.144754 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.144780 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.145178 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.145499 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.145655 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.145906 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.146271 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.146695 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.167887 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-89wdj"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.168562 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.169027 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.169563 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.171660 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.173146 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.167899 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfll6\" (UniqueName: \"kubernetes.io/projected/fd459946-dd06-42f3-896c-8ae5b7394b33-kube-api-access-rfll6\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbf9m\" (UID: \"fd459946-dd06-42f3-896c-8ae5b7394b33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.174956 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58b0ae05-55dc-4238-9163-c8e06974ffca-serving-cert\") pod \"route-controller-manager-6576b87f9c-h9dk9\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175022 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd459946-dd06-42f3-896c-8ae5b7394b33-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbf9m\" (UID: \"fd459946-dd06-42f3-896c-8ae5b7394b33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175048 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175078 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b66208b0-ad95-479b-8fae-c239c7d91fdb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175157 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175199 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd459946-dd06-42f3-896c-8ae5b7394b33-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbf9m\" (UID: \"fd459946-dd06-42f3-896c-8ae5b7394b33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175269 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b66208b0-ad95-479b-8fae-c239c7d91fdb-encryption-config\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175298 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-config\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175330 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58b0ae05-55dc-4238-9163-c8e06974ffca-client-ca\") pod \"route-controller-manager-6576b87f9c-h9dk9\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175357 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1f2720fe-737e-4562-9fc1-d0e52e0d99a2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8sdsb\" (UID: \"1f2720fe-737e-4562-9fc1-d0e52e0d99a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175408 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b66208b0-ad95-479b-8fae-c239c7d91fdb-audit-policies\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175476 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b66208b0-ad95-479b-8fae-c239c7d91fdb-audit-dir\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175520 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58b0ae05-55dc-4238-9163-c8e06974ffca-config\") pod \"route-controller-manager-6576b87f9c-h9dk9\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175583 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f2720fe-737e-4562-9fc1-d0e52e0d99a2-config\") pod \"machine-api-operator-5694c8668f-8sdsb\" (UID: \"1f2720fe-737e-4562-9fc1-d0e52e0d99a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175629 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b66208b0-ad95-479b-8fae-c239c7d91fdb-serving-cert\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175706 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b66208b0-ad95-479b-8fae-c239c7d91fdb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175735 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1f2720fe-737e-4562-9fc1-d0e52e0d99a2-images\") pod \"machine-api-operator-5694c8668f-8sdsb\" (UID: \"1f2720fe-737e-4562-9fc1-d0e52e0d99a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175761 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d85wc\" (UniqueName: \"kubernetes.io/projected/1f2720fe-737e-4562-9fc1-d0e52e0d99a2-kube-api-access-d85wc\") pod \"machine-api-operator-5694c8668f-8sdsb\" (UID: \"1f2720fe-737e-4562-9fc1-d0e52e0d99a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175812 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8490c962-38c0-4be5-b940-d4c4d08c8a55-serving-cert\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175830 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j72zg\" (UniqueName: \"kubernetes.io/projected/8490c962-38c0-4be5-b940-d4c4d08c8a55-kube-api-access-j72zg\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175849 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6ssz\" (UniqueName: \"kubernetes.io/projected/58b0ae05-55dc-4238-9163-c8e06974ffca-kube-api-access-w6ssz\") pod \"route-controller-manager-6576b87f9c-h9dk9\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175869 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-client-ca\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175895 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b66208b0-ad95-479b-8fae-c239c7d91fdb-etcd-client\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.175912 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k99j\" (UniqueName: \"kubernetes.io/projected/b66208b0-ad95-479b-8fae-c239c7d91fdb-kube-api-access-6k99j\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.176928 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-m4262"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.177015 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd459946-dd06-42f3-896c-8ae5b7394b33-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbf9m\" (UID: \"fd459946-dd06-42f3-896c-8ae5b7394b33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.177388 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8nftj"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.177718 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ph88v"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.180335 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b66208b0-ad95-479b-8fae-c239c7d91fdb-audit-policies\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.181078 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.181421 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-config\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.182097 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58b0ae05-55dc-4238-9163-c8e06974ffca-client-ca\") pod \"route-controller-manager-6576b87f9c-h9dk9\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.182483 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.183387 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b66208b0-ad95-479b-8fae-c239c7d91fdb-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.183496 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.183746 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58b0ae05-55dc-4238-9163-c8e06974ffca-serving-cert\") pod \"route-controller-manager-6576b87f9c-h9dk9\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.185470 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wmxwg"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.186157 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.186711 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.188681 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.189385 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.189753 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.192195 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.192332 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd459946-dd06-42f3-896c-8ae5b7394b33-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbf9m\" (UID: \"fd459946-dd06-42f3-896c-8ae5b7394b33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.192758 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.193082 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58b0ae05-55dc-4238-9163-c8e06974ffca-config\") pod \"route-controller-manager-6576b87f9c-h9dk9\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.193093 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-m4262" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.193111 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1f2720fe-737e-4562-9fc1-d0e52e0d99a2-images\") pod \"machine-api-operator-5694c8668f-8sdsb\" (UID: \"1f2720fe-737e-4562-9fc1-d0e52e0d99a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.193144 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b66208b0-ad95-479b-8fae-c239c7d91fdb-audit-dir\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.193275 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.193601 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b66208b0-ad95-479b-8fae-c239c7d91fdb-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.193845 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f2720fe-737e-4562-9fc1-d0e52e0d99a2-config\") pod \"machine-api-operator-5694c8668f-8sdsb\" (UID: \"1f2720fe-737e-4562-9fc1-d0e52e0d99a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.193935 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.197108 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.197901 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.198493 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-bnxzq"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.198920 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.199247 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.199468 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.200041 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1f2720fe-737e-4562-9fc1-d0e52e0d99a2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-8sdsb\" (UID: \"1f2720fe-737e-4562-9fc1-d0e52e0d99a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.201307 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8490c962-38c0-4be5-b940-d4c4d08c8a55-serving-cert\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.204923 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.205012 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.205474 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.206755 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-client-ca\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.210182 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.212252 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.212557 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.213402 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.222728 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b66208b0-ad95-479b-8fae-c239c7d91fdb-encryption-config\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.223053 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b66208b0-ad95-479b-8fae-c239c7d91fdb-etcd-client\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.223726 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.224606 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.222481 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b66208b0-ad95-479b-8fae-c239c7d91fdb-serving-cert\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.226665 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.231982 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rbhdx"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.232759 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.233955 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.234681 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.235604 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-zkjfk"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.238340 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.239913 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.240739 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.242736 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-8sdsb"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.246971 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.249157 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.252525 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.252537 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.253758 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2g96c"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.254807 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-28rrg"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.255871 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.256902 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wkldv"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.258043 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.259022 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-grw5p"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.260019 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-tf5g6"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.261192 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-fs4kj"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.262179 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.262540 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.264578 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.265789 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.266347 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.266987 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-89wdj"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.268171 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hgd88"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.269588 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.270760 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.272036 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ml4z6"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.273297 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kdt7l"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.275061 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.276366 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.276754 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14bd93bd-8f38-48c3-9508-6c1ab9e65628-audit-dir\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.276812 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7ec6444d-9ad0-4310-974a-d8525e16e950-machine-approver-tls\") pod \"machine-approver-56656f9798-h9rws\" (UID: \"7ec6444d-9ad0-4310-974a-d8525e16e950\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.276999 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-apiservice-cert\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277033 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3d93745e-ade6-45eb-b5e3-9b4270483319-etcd-client\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277056 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11c1150e-cb65-4c9d-98b6-6c36f5092307-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-lnrw7\" (UID: \"11c1150e-cb65-4c9d-98b6-6c36f5092307\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277080 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d93e50fc-a2f7-4ed5-ad71-7590b393c71f-serving-cert\") pod \"openshift-config-operator-7777fb866f-89wdj\" (UID: \"d93e50fc-a2f7-4ed5-ad71-7590b393c71f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277098 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c9f637d-2d43-4bb6-9db2-1831bd8e7573-serving-cert\") pod \"console-operator-58897d9998-tf5g6\" (UID: \"6c9f637d-2d43-4bb6-9db2-1831bd8e7573\") " pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277148 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-service-ca\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277190 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/560d2fc0-5e01-4d10-9499-72144de110f6-images\") pod \"machine-config-operator-74547568cd-fqmc8\" (UID: \"560d2fc0-5e01-4d10-9499-72144de110f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277229 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-image-import-ca\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277267 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5mtt\" (UniqueName: \"kubernetes.io/projected/7ec6444d-9ad0-4310-974a-d8525e16e950-kube-api-access-l5mtt\") pod \"machine-approver-56656f9798-h9rws\" (UID: \"7ec6444d-9ad0-4310-974a-d8525e16e950\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277338 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3c8ac74-bfd7-494b-9181-36832b3ffa98-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-779v8\" (UID: \"b3c8ac74-bfd7-494b-9181-36832b3ffa98\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277388 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b3c8ac74-bfd7-494b-9181-36832b3ffa98-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-779v8\" (UID: \"b3c8ac74-bfd7-494b-9181-36832b3ffa98\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277451 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b97e1699-6f5e-461d-b3e0-4afaef630174-config\") pod \"openshift-apiserver-operator-796bbdcf4f-cm2tk\" (UID: \"b97e1699-6f5e-461d-b3e0-4afaef630174\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277478 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/560d2fc0-5e01-4d10-9499-72144de110f6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fqmc8\" (UID: \"560d2fc0-5e01-4d10-9499-72144de110f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277629 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-config-volume\") pod \"collect-profiles-29414055-z8m4l\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277642 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277669 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-webhook-cert\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277697 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6c9f637d-2d43-4bb6-9db2-1831bd8e7573-trusted-ca\") pod \"console-operator-58897d9998-tf5g6\" (UID: \"6c9f637d-2d43-4bb6-9db2-1831bd8e7573\") " pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277725 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d93745e-ade6-45eb-b5e3-9b4270483319-audit-dir\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277759 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br2tk\" (UniqueName: \"kubernetes.io/projected/3d93745e-ade6-45eb-b5e3-9b4270483319-kube-api-access-br2tk\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277786 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2s7d\" (UniqueName: \"kubernetes.io/projected/14bd93bd-8f38-48c3-9508-6c1ab9e65628-kube-api-access-g2s7d\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277807 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9rrn\" (UniqueName: \"kubernetes.io/projected/7019146d-e522-4815-9642-9c9667aba409-kube-api-access-g9rrn\") pod \"dns-operator-744455d44c-ml4z6\" (UID: \"7019146d-e522-4815-9642-9c9667aba409\") " pod="openshift-dns-operator/dns-operator-744455d44c-ml4z6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277829 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0c0b640d-f71b-4c0e-8b92-e592722ccee5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-kvmzb\" (UID: \"0c0b640d-f71b-4c0e-8b92-e592722ccee5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277880 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277900 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7848703f-6d8c-4915-8f42-e6222f2a889a-service-ca\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277926 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e5550ab-55c3-4147-871f-3e49bf38ee55-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-hjcd2\" (UID: \"6e5550ab-55c3-4147-871f-3e49bf38ee55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.277966 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx5tf\" (UniqueName: \"kubernetes.io/projected/db1b9fc6-529d-4a5c-9ef9-b91b6f872225-kube-api-access-qx5tf\") pod \"service-ca-9c57cc56f-bnxzq\" (UID: \"db1b9fc6-529d-4a5c-9ef9-b91b6f872225\") " pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278110 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0c0b640d-f71b-4c0e-8b92-e592722ccee5-proxy-tls\") pod \"machine-config-controller-84d6567774-kvmzb\" (UID: \"0c0b640d-f71b-4c0e-8b92-e592722ccee5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278135 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/89aaae4b-99da-4820-b024-a013c629c63f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pxxxc\" (UID: \"89aaae4b-99da-4820-b024-a013c629c63f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278153 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278181 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7f0e5e19-36d5-4553-896d-c5057c8fe1c3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-m4262\" (UID: \"7f0e5e19-36d5-4553-896d-c5057c8fe1c3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m4262" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278213 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-audit-policies\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278233 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-ca\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278328 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp5tr\" (UniqueName: \"kubernetes.io/projected/7874756d-1da1-481a-be6e-df6034c4cac5-kube-api-access-xp5tr\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278359 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsmwg\" (UniqueName: \"kubernetes.io/projected/d93e50fc-a2f7-4ed5-ad71-7590b393c71f-kube-api-access-qsmwg\") pod \"openshift-config-operator-7777fb866f-89wdj\" (UID: \"d93e50fc-a2f7-4ed5-ad71-7590b393c71f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278384 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cefa13a6-8806-4b35-9083-df5aa7d45f56-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-snz6f\" (UID: \"cefa13a6-8806-4b35-9083-df5aa7d45f56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278413 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-client\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278434 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e188d23e-750e-4c02-9388-0c8bc4a2906d-profile-collector-cert\") pod \"catalog-operator-68c6474976-fzpv5\" (UID: \"e188d23e-750e-4c02-9388-0c8bc4a2906d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278456 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/560d2fc0-5e01-4d10-9499-72144de110f6-proxy-tls\") pod \"machine-config-operator-74547568cd-fqmc8\" (UID: \"560d2fc0-5e01-4d10-9499-72144de110f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278480 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llrpq\" (UniqueName: \"kubernetes.io/projected/b97e1699-6f5e-461d-b3e0-4afaef630174-kube-api-access-llrpq\") pod \"openshift-apiserver-operator-796bbdcf4f-cm2tk\" (UID: \"b97e1699-6f5e-461d-b3e0-4afaef630174\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278502 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0352a66-8ffd-4949-ba0f-d8e6cb933ee4-serving-cert\") pod \"service-ca-operator-777779d784-hgd88\" (UID: \"c0352a66-8ffd-4949-ba0f-d8e6cb933ee4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278527 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278546 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szwh5\" (UniqueName: \"kubernetes.io/projected/560d2fc0-5e01-4d10-9499-72144de110f6-kube-api-access-szwh5\") pod \"machine-config-operator-74547568cd-fqmc8\" (UID: \"560d2fc0-5e01-4d10-9499-72144de110f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278566 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7848703f-6d8c-4915-8f42-e6222f2a889a-console-serving-cert\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278595 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wgml\" (UniqueName: \"kubernetes.io/projected/c0352a66-8ffd-4949-ba0f-d8e6cb933ee4-kube-api-access-8wgml\") pod \"service-ca-operator-777779d784-hgd88\" (UID: \"c0352a66-8ffd-4949-ba0f-d8e6cb933ee4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278611 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4gk2\" (UniqueName: \"kubernetes.io/projected/f93fafb8-e1d5-4ef2-a443-b3cfe19b2bda-kube-api-access-f4gk2\") pod \"migrator-59844c95c7-28rrg\" (UID: \"f93fafb8-e1d5-4ef2-a443-b3cfe19b2bda\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-28rrg" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278629 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-config\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278649 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0352a66-8ffd-4949-ba0f-d8e6cb933ee4-config\") pod \"service-ca-operator-777779d784-hgd88\" (UID: \"c0352a66-8ffd-4949-ba0f-d8e6cb933ee4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278728 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c9f637d-2d43-4bb6-9db2-1831bd8e7573-config\") pod \"console-operator-58897d9998-tf5g6\" (UID: \"6c9f637d-2d43-4bb6-9db2-1831bd8e7573\") " pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278763 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e5550ab-55c3-4147-871f-3e49bf38ee55-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-hjcd2\" (UID: \"6e5550ab-55c3-4147-871f-3e49bf38ee55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278791 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77jvx\" (UniqueName: \"kubernetes.io/projected/1abfb759-04f9-4885-b150-82355f8cf866-kube-api-access-77jvx\") pod \"downloads-7954f5f757-grw5p\" (UID: \"1abfb759-04f9-4885-b150-82355f8cf866\") " pod="openshift-console/downloads-7954f5f757-grw5p" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278818 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/db1b9fc6-529d-4a5c-9ef9-b91b6f872225-signing-cabundle\") pod \"service-ca-9c57cc56f-bnxzq\" (UID: \"db1b9fc6-529d-4a5c-9ef9-b91b6f872225\") " pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278835 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-config\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278855 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-etcd-serving-ca\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278907 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278957 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cefa13a6-8806-4b35-9083-df5aa7d45f56-config\") pod \"kube-apiserver-operator-766d6c64bb-snz6f\" (UID: \"cefa13a6-8806-4b35-9083-df5aa7d45f56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278963 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8nftj"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.278979 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec6444d-9ad0-4310-974a-d8525e16e950-config\") pod \"machine-approver-56656f9798-h9rws\" (UID: \"7ec6444d-9ad0-4310-974a-d8525e16e950\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279001 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dn2r\" (UniqueName: \"kubernetes.io/projected/5ef5c158-a074-47df-a4bb-2e1cab760745-kube-api-access-8dn2r\") pod \"control-plane-machine-set-operator-78cbb6b69f-vhxkd\" (UID: \"5ef5c158-a074-47df-a4bb-2e1cab760745\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279028 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279076 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3c8ac74-bfd7-494b-9181-36832b3ffa98-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-779v8\" (UID: \"b3c8ac74-bfd7-494b-9181-36832b3ffa98\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279112 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c7m9\" (UniqueName: \"kubernetes.io/projected/931858b0-187e-441e-a961-80c7cf44d916-kube-api-access-8c7m9\") pod \"cluster-samples-operator-665b6dd947-cx7l7\" (UID: \"931858b0-187e-441e-a961-80c7cf44d916\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279137 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-secret-volume\") pod \"collect-profiles-29414055-z8m4l\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279192 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7848703f-6d8c-4915-8f42-e6222f2a889a-oauth-serving-cert\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279245 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5550ab-55c3-4147-871f-3e49bf38ee55-config\") pod \"kube-controller-manager-operator-78b949d7b-hjcd2\" (UID: \"6e5550ab-55c3-4147-871f-3e49bf38ee55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279282 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/931858b0-187e-441e-a961-80c7cf44d916-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-cx7l7\" (UID: \"931858b0-187e-441e-a961-80c7cf44d916\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279314 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhd9q\" (UniqueName: \"kubernetes.io/projected/e188d23e-750e-4c02-9388-0c8bc4a2906d-kube-api-access-rhd9q\") pod \"catalog-operator-68c6474976-fzpv5\" (UID: \"e188d23e-750e-4c02-9388-0c8bc4a2906d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279346 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-metrics-tls\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279380 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-trusted-ca\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279428 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279481 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-tmpfs\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279516 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11c1150e-cb65-4c9d-98b6-6c36f5092307-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-lnrw7\" (UID: \"11c1150e-cb65-4c9d-98b6-6c36f5092307\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279549 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-service-ca-bundle\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279590 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3d93745e-ade6-45eb-b5e3-9b4270483319-node-pullsecrets\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279637 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279672 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279706 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz5jv\" (UniqueName: \"kubernetes.io/projected/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-kube-api-access-kz5jv\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279746 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdf2c613-8426-4c4e-8cfe-bc015e17e47e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5dsb\" (UID: \"cdf2c613-8426-4c4e-8cfe-bc015e17e47e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279808 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc5ss\" (UniqueName: \"kubernetes.io/projected/6c9f637d-2d43-4bb6-9db2-1831bd8e7573-kube-api-access-zc5ss\") pod \"console-operator-58897d9998-tf5g6\" (UID: \"6c9f637d-2d43-4bb6-9db2-1831bd8e7573\") " pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279854 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtjkn\" (UniqueName: \"kubernetes.io/projected/7848703f-6d8c-4915-8f42-e6222f2a889a-kube-api-access-xtjkn\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279905 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4f4s\" (UniqueName: \"kubernetes.io/projected/b3c8ac74-bfd7-494b-9181-36832b3ffa98-kube-api-access-j4f4s\") pod \"cluster-image-registry-operator-dc59b4c8b-779v8\" (UID: \"b3c8ac74-bfd7-494b-9181-36832b3ffa98\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279930 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279960 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.279983 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjk27\" (UniqueName: \"kubernetes.io/projected/7f0e5e19-36d5-4553-896d-c5057c8fe1c3-kube-api-access-zjk27\") pod \"multus-admission-controller-857f4d67dd-m4262\" (UID: \"7f0e5e19-36d5-4553-896d-c5057c8fe1c3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m4262" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280020 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280079 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d93745e-ade6-45eb-b5e3-9b4270483319-serving-cert\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280138 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/89aaae4b-99da-4820-b024-a013c629c63f-srv-cert\") pod \"olm-operator-6b444d44fb-pxxxc\" (UID: \"89aaae4b-99da-4820-b024-a013c629c63f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280165 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cdf2c613-8426-4c4e-8cfe-bc015e17e47e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5dsb\" (UID: \"cdf2c613-8426-4c4e-8cfe-bc015e17e47e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280183 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e188d23e-750e-4c02-9388-0c8bc4a2906d-srv-cert\") pod \"catalog-operator-68c6474976-fzpv5\" (UID: \"e188d23e-750e-4c02-9388-0c8bc4a2906d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280222 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280244 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxbwf\" (UniqueName: \"kubernetes.io/projected/89aaae4b-99da-4820-b024-a013c629c63f-kube-api-access-fxbwf\") pod \"olm-operator-6b444d44fb-pxxxc\" (UID: \"89aaae4b-99da-4820-b024-a013c629c63f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280261 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3d93745e-ade6-45eb-b5e3-9b4270483319-encryption-config\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280278 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7874756d-1da1-481a-be6e-df6034c4cac5-serving-cert\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280298 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7ec6444d-9ad0-4310-974a-d8525e16e950-auth-proxy-config\") pod \"machine-approver-56656f9798-h9rws\" (UID: \"7ec6444d-9ad0-4310-974a-d8525e16e950\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280349 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7848703f-6d8c-4915-8f42-e6222f2a889a-console-oauth-config\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280385 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b97e1699-6f5e-461d-b3e0-4afaef630174-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-cm2tk\" (UID: \"b97e1699-6f5e-461d-b3e0-4afaef630174\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280421 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pnm9\" (UniqueName: \"kubernetes.io/projected/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-kube-api-access-8pnm9\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280445 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz568\" (UniqueName: \"kubernetes.io/projected/11c1150e-cb65-4c9d-98b6-6c36f5092307-kube-api-access-bz568\") pod \"kube-storage-version-migrator-operator-b67b599dd-lnrw7\" (UID: \"11c1150e-cb65-4c9d-98b6-6c36f5092307\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280473 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280494 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280519 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ph88v"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280525 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cdf2c613-8426-4c4e-8cfe-bc015e17e47e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5dsb\" (UID: \"cdf2c613-8426-4c4e-8cfe-bc015e17e47e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280556 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/db1b9fc6-529d-4a5c-9ef9-b91b6f872225-signing-key\") pod \"service-ca-9c57cc56f-bnxzq\" (UID: \"db1b9fc6-529d-4a5c-9ef9-b91b6f872225\") " pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280579 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nfln\" (UniqueName: \"kubernetes.io/projected/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-kube-api-access-7nfln\") pod \"collect-profiles-29414055-z8m4l\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280602 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq479\" (UniqueName: \"kubernetes.io/projected/3e8ecae4-2aee-4125-a16c-d4f340d840de-kube-api-access-qq479\") pod \"package-server-manager-789f6589d5-ws8hv\" (UID: \"3e8ecae4-2aee-4125-a16c-d4f340d840de\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280623 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5ef5c158-a074-47df-a4bb-2e1cab760745-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vhxkd\" (UID: \"5ef5c158-a074-47df-a4bb-2e1cab760745\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280668 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3e8ecae4-2aee-4125-a16c-d4f340d840de-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ws8hv\" (UID: \"3e8ecae4-2aee-4125-a16c-d4f340d840de\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280702 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7019146d-e522-4815-9642-9c9667aba409-metrics-tls\") pod \"dns-operator-744455d44c-ml4z6\" (UID: \"7019146d-e522-4815-9642-9c9667aba409\") " pod="openshift-dns-operator/dns-operator-744455d44c-ml4z6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280750 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7848703f-6d8c-4915-8f42-e6222f2a889a-console-config\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280775 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-config\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280802 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-serving-cert\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280827 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5wbr\" (UniqueName: \"kubernetes.io/projected/0c0b640d-f71b-4c0e-8b92-e592722ccee5-kube-api-access-n5wbr\") pod \"machine-config-controller-84d6567774-kvmzb\" (UID: \"0c0b640d-f71b-4c0e-8b92-e592722ccee5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280850 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d93e50fc-a2f7-4ed5-ad71-7590b393c71f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-89wdj\" (UID: \"d93e50fc-a2f7-4ed5-ad71-7590b393c71f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280873 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-audit\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280899 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfztz\" (UniqueName: \"kubernetes.io/projected/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-kube-api-access-xfztz\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280922 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7848703f-6d8c-4915-8f42-e6222f2a889a-trusted-ca-bundle\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.280944 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cefa13a6-8806-4b35-9083-df5aa7d45f56-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-snz6f\" (UID: \"cefa13a6-8806-4b35-9083-df5aa7d45f56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.282294 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.283744 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.284983 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-rqdpv"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.286397 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-6jx2n"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.286479 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.287039 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-6jx2n" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.287371 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.287429 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-m4262"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.298298 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-bnxzq"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.301335 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.302665 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.304209 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-fs4kj"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.306464 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.313418 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.315270 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.316969 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5bgm4"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.318744 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-rqdpv"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.321056 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wmxwg"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.322377 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-6jx2n"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.323435 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rbhdx"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.324455 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-xrlzc"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.324986 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xrlzc" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.326253 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.366977 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381636 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e188d23e-750e-4c02-9388-0c8bc4a2906d-srv-cert\") pod \"catalog-operator-68c6474976-fzpv5\" (UID: \"e188d23e-750e-4c02-9388-0c8bc4a2906d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381672 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381692 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381713 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d93745e-ade6-45eb-b5e3-9b4270483319-serving-cert\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381728 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/89aaae4b-99da-4820-b024-a013c629c63f-srv-cert\") pod \"olm-operator-6b444d44fb-pxxxc\" (UID: \"89aaae4b-99da-4820-b024-a013c629c63f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381742 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cdf2c613-8426-4c4e-8cfe-bc015e17e47e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5dsb\" (UID: \"cdf2c613-8426-4c4e-8cfe-bc015e17e47e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381757 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxbwf\" (UniqueName: \"kubernetes.io/projected/89aaae4b-99da-4820-b024-a013c629c63f-kube-api-access-fxbwf\") pod \"olm-operator-6b444d44fb-pxxxc\" (UID: \"89aaae4b-99da-4820-b024-a013c629c63f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381771 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3d93745e-ade6-45eb-b5e3-9b4270483319-encryption-config\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381785 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7874756d-1da1-481a-be6e-df6034c4cac5-serving-cert\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381800 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7ec6444d-9ad0-4310-974a-d8525e16e950-auth-proxy-config\") pod \"machine-approver-56656f9798-h9rws\" (UID: \"7ec6444d-9ad0-4310-974a-d8525e16e950\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381816 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b97e1699-6f5e-461d-b3e0-4afaef630174-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-cm2tk\" (UID: \"b97e1699-6f5e-461d-b3e0-4afaef630174\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381835 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7848703f-6d8c-4915-8f42-e6222f2a889a-console-oauth-config\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381851 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381866 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pnm9\" (UniqueName: \"kubernetes.io/projected/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-kube-api-access-8pnm9\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381883 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz568\" (UniqueName: \"kubernetes.io/projected/11c1150e-cb65-4c9d-98b6-6c36f5092307-kube-api-access-bz568\") pod \"kube-storage-version-migrator-operator-b67b599dd-lnrw7\" (UID: \"11c1150e-cb65-4c9d-98b6-6c36f5092307\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381900 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq479\" (UniqueName: \"kubernetes.io/projected/3e8ecae4-2aee-4125-a16c-d4f340d840de-kube-api-access-qq479\") pod \"package-server-manager-789f6589d5-ws8hv\" (UID: \"3e8ecae4-2aee-4125-a16c-d4f340d840de\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381919 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5ef5c158-a074-47df-a4bb-2e1cab760745-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vhxkd\" (UID: \"5ef5c158-a074-47df-a4bb-2e1cab760745\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381934 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381947 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cdf2c613-8426-4c4e-8cfe-bc015e17e47e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5dsb\" (UID: \"cdf2c613-8426-4c4e-8cfe-bc015e17e47e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381972 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/db1b9fc6-529d-4a5c-9ef9-b91b6f872225-signing-key\") pod \"service-ca-9c57cc56f-bnxzq\" (UID: \"db1b9fc6-529d-4a5c-9ef9-b91b6f872225\") " pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.381988 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nfln\" (UniqueName: \"kubernetes.io/projected/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-kube-api-access-7nfln\") pod \"collect-profiles-29414055-z8m4l\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382003 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3e8ecae4-2aee-4125-a16c-d4f340d840de-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ws8hv\" (UID: \"3e8ecae4-2aee-4125-a16c-d4f340d840de\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382019 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7019146d-e522-4815-9642-9c9667aba409-metrics-tls\") pod \"dns-operator-744455d44c-ml4z6\" (UID: \"7019146d-e522-4815-9642-9c9667aba409\") " pod="openshift-dns-operator/dns-operator-744455d44c-ml4z6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382034 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7848703f-6d8c-4915-8f42-e6222f2a889a-console-config\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382050 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-config\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382068 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-serving-cert\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382100 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5wbr\" (UniqueName: \"kubernetes.io/projected/0c0b640d-f71b-4c0e-8b92-e592722ccee5-kube-api-access-n5wbr\") pod \"machine-config-controller-84d6567774-kvmzb\" (UID: \"0c0b640d-f71b-4c0e-8b92-e592722ccee5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382118 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d93e50fc-a2f7-4ed5-ad71-7590b393c71f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-89wdj\" (UID: \"d93e50fc-a2f7-4ed5-ad71-7590b393c71f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382135 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-audit\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382165 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfztz\" (UniqueName: \"kubernetes.io/projected/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-kube-api-access-xfztz\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382183 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7848703f-6d8c-4915-8f42-e6222f2a889a-trusted-ca-bundle\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382222 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cefa13a6-8806-4b35-9083-df5aa7d45f56-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-snz6f\" (UID: \"cefa13a6-8806-4b35-9083-df5aa7d45f56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382240 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14bd93bd-8f38-48c3-9508-6c1ab9e65628-audit-dir\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382256 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7ec6444d-9ad0-4310-974a-d8525e16e950-machine-approver-tls\") pod \"machine-approver-56656f9798-h9rws\" (UID: \"7ec6444d-9ad0-4310-974a-d8525e16e950\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382272 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-apiservice-cert\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382289 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3d93745e-ade6-45eb-b5e3-9b4270483319-etcd-client\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382305 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11c1150e-cb65-4c9d-98b6-6c36f5092307-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-lnrw7\" (UID: \"11c1150e-cb65-4c9d-98b6-6c36f5092307\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382321 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d93e50fc-a2f7-4ed5-ad71-7590b393c71f-serving-cert\") pod \"openshift-config-operator-7777fb866f-89wdj\" (UID: \"d93e50fc-a2f7-4ed5-ad71-7590b393c71f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382339 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c9f637d-2d43-4bb6-9db2-1831bd8e7573-serving-cert\") pod \"console-operator-58897d9998-tf5g6\" (UID: \"6c9f637d-2d43-4bb6-9db2-1831bd8e7573\") " pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382367 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/560d2fc0-5e01-4d10-9499-72144de110f6-images\") pod \"machine-config-operator-74547568cd-fqmc8\" (UID: \"560d2fc0-5e01-4d10-9499-72144de110f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382386 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-image-import-ca\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382404 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-service-ca\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382421 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3c8ac74-bfd7-494b-9181-36832b3ffa98-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-779v8\" (UID: \"b3c8ac74-bfd7-494b-9181-36832b3ffa98\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382438 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b3c8ac74-bfd7-494b-9181-36832b3ffa98-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-779v8\" (UID: \"b3c8ac74-bfd7-494b-9181-36832b3ffa98\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382452 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5mtt\" (UniqueName: \"kubernetes.io/projected/7ec6444d-9ad0-4310-974a-d8525e16e950-kube-api-access-l5mtt\") pod \"machine-approver-56656f9798-h9rws\" (UID: \"7ec6444d-9ad0-4310-974a-d8525e16e950\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382476 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b97e1699-6f5e-461d-b3e0-4afaef630174-config\") pod \"openshift-apiserver-operator-796bbdcf4f-cm2tk\" (UID: \"b97e1699-6f5e-461d-b3e0-4afaef630174\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382498 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/560d2fc0-5e01-4d10-9499-72144de110f6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fqmc8\" (UID: \"560d2fc0-5e01-4d10-9499-72144de110f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382554 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-webhook-cert\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382570 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-config-volume\") pod \"collect-profiles-29414055-z8m4l\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382592 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2s7d\" (UniqueName: \"kubernetes.io/projected/14bd93bd-8f38-48c3-9508-6c1ab9e65628-kube-api-access-g2s7d\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382610 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9rrn\" (UniqueName: \"kubernetes.io/projected/7019146d-e522-4815-9642-9c9667aba409-kube-api-access-g9rrn\") pod \"dns-operator-744455d44c-ml4z6\" (UID: \"7019146d-e522-4815-9642-9c9667aba409\") " pod="openshift-dns-operator/dns-operator-744455d44c-ml4z6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382628 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6c9f637d-2d43-4bb6-9db2-1831bd8e7573-trusted-ca\") pod \"console-operator-58897d9998-tf5g6\" (UID: \"6c9f637d-2d43-4bb6-9db2-1831bd8e7573\") " pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382646 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d93745e-ade6-45eb-b5e3-9b4270483319-audit-dir\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382662 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br2tk\" (UniqueName: \"kubernetes.io/projected/3d93745e-ade6-45eb-b5e3-9b4270483319-kube-api-access-br2tk\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382685 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0c0b640d-f71b-4c0e-8b92-e592722ccee5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-kvmzb\" (UID: \"0c0b640d-f71b-4c0e-8b92-e592722ccee5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382702 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx5tf\" (UniqueName: \"kubernetes.io/projected/db1b9fc6-529d-4a5c-9ef9-b91b6f872225-kube-api-access-qx5tf\") pod \"service-ca-9c57cc56f-bnxzq\" (UID: \"db1b9fc6-529d-4a5c-9ef9-b91b6f872225\") " pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382721 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382738 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7848703f-6d8c-4915-8f42-e6222f2a889a-service-ca\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382755 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e5550ab-55c3-4147-871f-3e49bf38ee55-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-hjcd2\" (UID: \"6e5550ab-55c3-4147-871f-3e49bf38ee55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382775 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382807 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0c0b640d-f71b-4c0e-8b92-e592722ccee5-proxy-tls\") pod \"machine-config-controller-84d6567774-kvmzb\" (UID: \"0c0b640d-f71b-4c0e-8b92-e592722ccee5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382841 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/89aaae4b-99da-4820-b024-a013c629c63f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pxxxc\" (UID: \"89aaae4b-99da-4820-b024-a013c629c63f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382878 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-audit-policies\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382901 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7f0e5e19-36d5-4553-896d-c5057c8fe1c3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-m4262\" (UID: \"7f0e5e19-36d5-4553-896d-c5057c8fe1c3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m4262" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382926 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-ca\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382949 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp5tr\" (UniqueName: \"kubernetes.io/projected/7874756d-1da1-481a-be6e-df6034c4cac5-kube-api-access-xp5tr\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382973 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsmwg\" (UniqueName: \"kubernetes.io/projected/d93e50fc-a2f7-4ed5-ad71-7590b393c71f-kube-api-access-qsmwg\") pod \"openshift-config-operator-7777fb866f-89wdj\" (UID: \"d93e50fc-a2f7-4ed5-ad71-7590b393c71f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.382998 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cefa13a6-8806-4b35-9083-df5aa7d45f56-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-snz6f\" (UID: \"cefa13a6-8806-4b35-9083-df5aa7d45f56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.383020 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/560d2fc0-5e01-4d10-9499-72144de110f6-proxy-tls\") pod \"machine-config-operator-74547568cd-fqmc8\" (UID: \"560d2fc0-5e01-4d10-9499-72144de110f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.383042 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llrpq\" (UniqueName: \"kubernetes.io/projected/b97e1699-6f5e-461d-b3e0-4afaef630174-kube-api-access-llrpq\") pod \"openshift-apiserver-operator-796bbdcf4f-cm2tk\" (UID: \"b97e1699-6f5e-461d-b3e0-4afaef630174\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.383064 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-client\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.383804 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d93e50fc-a2f7-4ed5-ad71-7590b393c71f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-89wdj\" (UID: \"d93e50fc-a2f7-4ed5-ad71-7590b393c71f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.383856 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/560d2fc0-5e01-4d10-9499-72144de110f6-images\") pod \"machine-config-operator-74547568cd-fqmc8\" (UID: \"560d2fc0-5e01-4d10-9499-72144de110f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.384308 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7848703f-6d8c-4915-8f42-e6222f2a889a-console-config\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.384352 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e188d23e-750e-4c02-9388-0c8bc4a2906d-profile-collector-cert\") pod \"catalog-operator-68c6474976-fzpv5\" (UID: \"e188d23e-750e-4c02-9388-0c8bc4a2906d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.384403 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0352a66-8ffd-4949-ba0f-d8e6cb933ee4-serving-cert\") pod \"service-ca-operator-777779d784-hgd88\" (UID: \"c0352a66-8ffd-4949-ba0f-d8e6cb933ee4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.384432 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.385249 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-image-import-ca\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.385488 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3d93745e-ade6-45eb-b5e3-9b4270483319-audit-dir\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.385755 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/560d2fc0-5e01-4d10-9499-72144de110f6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fqmc8\" (UID: \"560d2fc0-5e01-4d10-9499-72144de110f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.386364 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14bd93bd-8f38-48c3-9508-6c1ab9e65628-audit-dir\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.386369 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b97e1699-6f5e-461d-b3e0-4afaef630174-config\") pod \"openshift-apiserver-operator-796bbdcf4f-cm2tk\" (UID: \"b97e1699-6f5e-461d-b3e0-4afaef630174\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.386769 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7848703f-6d8c-4915-8f42-e6222f2a889a-service-ca\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.386862 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szwh5\" (UniqueName: \"kubernetes.io/projected/560d2fc0-5e01-4d10-9499-72144de110f6-kube-api-access-szwh5\") pod \"machine-config-operator-74547568cd-fqmc8\" (UID: \"560d2fc0-5e01-4d10-9499-72144de110f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.386890 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7848703f-6d8c-4915-8f42-e6222f2a889a-console-serving-cert\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.386911 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wgml\" (UniqueName: \"kubernetes.io/projected/c0352a66-8ffd-4949-ba0f-d8e6cb933ee4-kube-api-access-8wgml\") pod \"service-ca-operator-777779d784-hgd88\" (UID: \"c0352a66-8ffd-4949-ba0f-d8e6cb933ee4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.386948 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4gk2\" (UniqueName: \"kubernetes.io/projected/f93fafb8-e1d5-4ef2-a443-b3cfe19b2bda-kube-api-access-f4gk2\") pod \"migrator-59844c95c7-28rrg\" (UID: \"f93fafb8-e1d5-4ef2-a443-b3cfe19b2bda\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-28rrg" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.386967 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-config\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.386987 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0352a66-8ffd-4949-ba0f-d8e6cb933ee4-config\") pod \"service-ca-operator-777779d784-hgd88\" (UID: \"c0352a66-8ffd-4949-ba0f-d8e6cb933ee4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387050 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77jvx\" (UniqueName: \"kubernetes.io/projected/1abfb759-04f9-4885-b150-82355f8cf866-kube-api-access-77jvx\") pod \"downloads-7954f5f757-grw5p\" (UID: \"1abfb759-04f9-4885-b150-82355f8cf866\") " pod="openshift-console/downloads-7954f5f757-grw5p" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387066 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c9f637d-2d43-4bb6-9db2-1831bd8e7573-config\") pod \"console-operator-58897d9998-tf5g6\" (UID: \"6c9f637d-2d43-4bb6-9db2-1831bd8e7573\") " pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387084 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e5550ab-55c3-4147-871f-3e49bf38ee55-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-hjcd2\" (UID: \"6e5550ab-55c3-4147-871f-3e49bf38ee55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387109 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/db1b9fc6-529d-4a5c-9ef9-b91b6f872225-signing-cabundle\") pod \"service-ca-9c57cc56f-bnxzq\" (UID: \"db1b9fc6-529d-4a5c-9ef9-b91b6f872225\") " pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387126 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387145 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cefa13a6-8806-4b35-9083-df5aa7d45f56-config\") pod \"kube-apiserver-operator-766d6c64bb-snz6f\" (UID: \"cefa13a6-8806-4b35-9083-df5aa7d45f56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387161 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-config\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387178 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-etcd-serving-ca\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387223 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3c8ac74-bfd7-494b-9181-36832b3ffa98-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-779v8\" (UID: \"b3c8ac74-bfd7-494b-9181-36832b3ffa98\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387257 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec6444d-9ad0-4310-974a-d8525e16e950-config\") pod \"machine-approver-56656f9798-h9rws\" (UID: \"7ec6444d-9ad0-4310-974a-d8525e16e950\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387284 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dn2r\" (UniqueName: \"kubernetes.io/projected/5ef5c158-a074-47df-a4bb-2e1cab760745-kube-api-access-8dn2r\") pod \"control-plane-machine-set-operator-78cbb6b69f-vhxkd\" (UID: \"5ef5c158-a074-47df-a4bb-2e1cab760745\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387313 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387346 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c7m9\" (UniqueName: \"kubernetes.io/projected/931858b0-187e-441e-a961-80c7cf44d916-kube-api-access-8c7m9\") pod \"cluster-samples-operator-665b6dd947-cx7l7\" (UID: \"931858b0-187e-441e-a961-80c7cf44d916\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387373 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-secret-volume\") pod \"collect-profiles-29414055-z8m4l\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387401 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7848703f-6d8c-4915-8f42-e6222f2a889a-oauth-serving-cert\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387420 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5550ab-55c3-4147-871f-3e49bf38ee55-config\") pod \"kube-controller-manager-operator-78b949d7b-hjcd2\" (UID: \"6e5550ab-55c3-4147-871f-3e49bf38ee55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387436 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/931858b0-187e-441e-a961-80c7cf44d916-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-cx7l7\" (UID: \"931858b0-187e-441e-a961-80c7cf44d916\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387453 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhd9q\" (UniqueName: \"kubernetes.io/projected/e188d23e-750e-4c02-9388-0c8bc4a2906d-kube-api-access-rhd9q\") pod \"catalog-operator-68c6474976-fzpv5\" (UID: \"e188d23e-750e-4c02-9388-0c8bc4a2906d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387481 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-metrics-tls\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387501 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-trusted-ca\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387565 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387590 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-tmpfs\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387657 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-audit-policies\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387652 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11c1150e-cb65-4c9d-98b6-6c36f5092307-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-lnrw7\" (UID: \"11c1150e-cb65-4c9d-98b6-6c36f5092307\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387743 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-service-ca-bundle\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387780 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7848703f-6d8c-4915-8f42-e6222f2a889a-trusted-ca-bundle\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387919 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6c9f637d-2d43-4bb6-9db2-1831bd8e7573-trusted-ca\") pod \"console-operator-58897d9998-tf5g6\" (UID: \"6c9f637d-2d43-4bb6-9db2-1831bd8e7573\") " pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387969 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-audit\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387974 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.387761 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3d93745e-ade6-45eb-b5e3-9b4270483319-node-pullsecrets\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.388128 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.388175 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.388182 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7848703f-6d8c-4915-8f42-e6222f2a889a-console-oauth-config\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.388221 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz5jv\" (UniqueName: \"kubernetes.io/projected/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-kube-api-access-kz5jv\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.388238 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3d93745e-ade6-45eb-b5e3-9b4270483319-encryption-config\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.388291 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-trusted-ca-bundle\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.388245 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdf2c613-8426-4c4e-8cfe-bc015e17e47e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5dsb\" (UID: \"cdf2c613-8426-4c4e-8cfe-bc015e17e47e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.388596 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7ec6444d-9ad0-4310-974a-d8525e16e950-auth-proxy-config\") pod \"machine-approver-56656f9798-h9rws\" (UID: \"7ec6444d-9ad0-4310-974a-d8525e16e950\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.388726 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-etcd-serving-ca\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.388650 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4f4s\" (UniqueName: \"kubernetes.io/projected/b3c8ac74-bfd7-494b-9181-36832b3ffa98-kube-api-access-j4f4s\") pod \"cluster-image-registry-operator-dc59b4c8b-779v8\" (UID: \"b3c8ac74-bfd7-494b-9181-36832b3ffa98\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.388899 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc5ss\" (UniqueName: \"kubernetes.io/projected/6c9f637d-2d43-4bb6-9db2-1831bd8e7573-kube-api-access-zc5ss\") pod \"console-operator-58897d9998-tf5g6\" (UID: \"6c9f637d-2d43-4bb6-9db2-1831bd8e7573\") " pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.388939 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtjkn\" (UniqueName: \"kubernetes.io/projected/7848703f-6d8c-4915-8f42-e6222f2a889a-kube-api-access-xtjkn\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.389080 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.389109 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.389261 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjk27\" (UniqueName: \"kubernetes.io/projected/7f0e5e19-36d5-4553-896d-c5057c8fe1c3-kube-api-access-zjk27\") pod \"multus-admission-controller-857f4d67dd-m4262\" (UID: \"7f0e5e19-36d5-4553-896d-c5057c8fe1c3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m4262" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.389346 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c0352a66-8ffd-4949-ba0f-d8e6cb933ee4-config\") pod \"service-ca-operator-777779d784-hgd88\" (UID: \"c0352a66-8ffd-4949-ba0f-d8e6cb933ee4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.389655 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.389834 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b3c8ac74-bfd7-494b-9181-36832b3ffa98-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-779v8\" (UID: \"b3c8ac74-bfd7-494b-9181-36832b3ffa98\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.390041 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d93745e-ade6-45eb-b5e3-9b4270483319-config\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.390271 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.390288 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cefa13a6-8806-4b35-9083-df5aa7d45f56-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-snz6f\" (UID: \"cefa13a6-8806-4b35-9083-df5aa7d45f56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.390828 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/3d93745e-ade6-45eb-b5e3-9b4270483319-node-pullsecrets\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.390899 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c9f637d-2d43-4bb6-9db2-1831bd8e7573-config\") pod \"console-operator-58897d9998-tf5g6\" (UID: \"6c9f637d-2d43-4bb6-9db2-1831bd8e7573\") " pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.391244 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-tmpfs\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.391488 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7848703f-6d8c-4915-8f42-e6222f2a889a-oauth-serving-cert\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.391653 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cefa13a6-8806-4b35-9083-df5aa7d45f56-config\") pod \"kube-apiserver-operator-766d6c64bb-snz6f\" (UID: \"cefa13a6-8806-4b35-9083-df5aa7d45f56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.392083 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.393497 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ec6444d-9ad0-4310-974a-d8525e16e950-config\") pod \"machine-approver-56656f9798-h9rws\" (UID: \"7ec6444d-9ad0-4310-974a-d8525e16e950\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.393994 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0c0b640d-f71b-4c0e-8b92-e592722ccee5-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-kvmzb\" (UID: \"0c0b640d-f71b-4c0e-8b92-e592722ccee5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.395732 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/560d2fc0-5e01-4d10-9499-72144de110f6-proxy-tls\") pod \"machine-config-operator-74547568cd-fqmc8\" (UID: \"560d2fc0-5e01-4d10-9499-72144de110f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.396130 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d93745e-ade6-45eb-b5e3-9b4270483319-serving-cert\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.396213 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.396448 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5ef5c158-a074-47df-a4bb-2e1cab760745-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-vhxkd\" (UID: \"5ef5c158-a074-47df-a4bb-2e1cab760745\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.396619 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b3c8ac74-bfd7-494b-9181-36832b3ffa98-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-779v8\" (UID: \"b3c8ac74-bfd7-494b-9181-36832b3ffa98\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.397002 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.397180 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3d93745e-ade6-45eb-b5e3-9b4270483319-etcd-client\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.397699 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b97e1699-6f5e-461d-b3e0-4afaef630174-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-cm2tk\" (UID: \"b97e1699-6f5e-461d-b3e0-4afaef630174\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.397977 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.398224 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.398472 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7848703f-6d8c-4915-8f42-e6222f2a889a-console-serving-cert\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.398836 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7019146d-e522-4815-9642-9c9667aba409-metrics-tls\") pod \"dns-operator-744455d44c-ml4z6\" (UID: \"7019146d-e522-4815-9642-9c9667aba409\") " pod="openshift-dns-operator/dns-operator-744455d44c-ml4z6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.398972 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.399107 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.399121 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c0352a66-8ffd-4949-ba0f-d8e6cb933ee4-serving-cert\") pod \"service-ca-operator-777779d784-hgd88\" (UID: \"c0352a66-8ffd-4949-ba0f-d8e6cb933ee4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.399171 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c9f637d-2d43-4bb6-9db2-1831bd8e7573-serving-cert\") pod \"console-operator-58897d9998-tf5g6\" (UID: \"6c9f637d-2d43-4bb6-9db2-1831bd8e7573\") " pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.399555 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.399672 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/931858b0-187e-441e-a961-80c7cf44d916-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-cx7l7\" (UID: \"931858b0-187e-441e-a961-80c7cf44d916\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.399698 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfll6\" (UniqueName: \"kubernetes.io/projected/fd459946-dd06-42f3-896c-8ae5b7394b33-kube-api-access-rfll6\") pod \"openshift-controller-manager-operator-756b6f6bc6-jbf9m\" (UID: \"fd459946-dd06-42f3-896c-8ae5b7394b33\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.402608 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/7ec6444d-9ad0-4310-974a-d8525e16e950-machine-approver-tls\") pod \"machine-approver-56656f9798-h9rws\" (UID: \"7ec6444d-9ad0-4310-974a-d8525e16e950\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.406878 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.429994 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.446382 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.450313 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d93e50fc-a2f7-4ed5-ad71-7590b393c71f-serving-cert\") pod \"openshift-config-operator-7777fb866f-89wdj\" (UID: \"d93e50fc-a2f7-4ed5-ad71-7590b393c71f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.467242 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.503175 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k99j\" (UniqueName: \"kubernetes.io/projected/b66208b0-ad95-479b-8fae-c239c7d91fdb-kube-api-access-6k99j\") pod \"apiserver-7bbb656c7d-4xwrm\" (UID: \"b66208b0-ad95-479b-8fae-c239c7d91fdb\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.506957 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.526883 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.546557 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.556074 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11c1150e-cb65-4c9d-98b6-6c36f5092307-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-lnrw7\" (UID: \"11c1150e-cb65-4c9d-98b6-6c36f5092307\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.566475 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.567982 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11c1150e-cb65-4c9d-98b6-6c36f5092307-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-lnrw7\" (UID: \"11c1150e-cb65-4c9d-98b6-6c36f5092307\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.587510 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.597602 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3e8ecae4-2aee-4125-a16c-d4f340d840de-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ws8hv\" (UID: \"3e8ecae4-2aee-4125-a16c-d4f340d840de\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.607368 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.626423 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.647542 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.664744 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.666679 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.681116 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.705430 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6ssz\" (UniqueName: \"kubernetes.io/projected/58b0ae05-55dc-4238-9163-c8e06974ffca-kube-api-access-w6ssz\") pod \"route-controller-manager-6576b87f9c-h9dk9\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.721259 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j72zg\" (UniqueName: \"kubernetes.io/projected/8490c962-38c0-4be5-b940-d4c4d08c8a55-kube-api-access-j72zg\") pod \"controller-manager-879f6c89f-2g96c\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.726478 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.732194 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e188d23e-750e-4c02-9388-0c8bc4a2906d-profile-collector-cert\") pod \"catalog-operator-68c6474976-fzpv5\" (UID: \"e188d23e-750e-4c02-9388-0c8bc4a2906d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.734283 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/89aaae4b-99da-4820-b024-a013c629c63f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-pxxxc\" (UID: \"89aaae4b-99da-4820-b024-a013c629c63f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.734416 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-secret-volume\") pod \"collect-profiles-29414055-z8m4l\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.735853 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.751287 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.767110 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.787257 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.798813 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/89aaae4b-99da-4820-b024-a013c629c63f-srv-cert\") pod \"olm-operator-6b444d44fb-pxxxc\" (UID: \"89aaae4b-99da-4820-b024-a013c629c63f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.808309 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.827372 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.836534 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cdf2c613-8426-4c4e-8cfe-bc015e17e47e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5dsb\" (UID: \"cdf2c613-8426-4c4e-8cfe-bc015e17e47e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.846931 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.849443 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdf2c613-8426-4c4e-8cfe-bc015e17e47e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5dsb\" (UID: \"cdf2c613-8426-4c4e-8cfe-bc015e17e47e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.867762 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.867911 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.869072 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm"] Dec 04 10:18:17 crc kubenswrapper[4943]: W1204 10:18:17.876475 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd459946_dd06_42f3_896c_8ae5b7394b33.slice/crio-98269efd2678af80f7b3942a937a01a506044348ccb6ed7161bd46fc93a7109f WatchSource:0}: Error finding container 98269efd2678af80f7b3942a937a01a506044348ccb6ed7161bd46fc93a7109f: Status 404 returned error can't find the container with id 98269efd2678af80f7b3942a937a01a506044348ccb6ed7161bd46fc93a7109f Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.887021 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.910581 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.927390 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9"] Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.928193 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.936727 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.939828 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7f0e5e19-36d5-4553-896d-c5057c8fe1c3-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-m4262\" (UID: \"7f0e5e19-36d5-4553-896d-c5057c8fe1c3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m4262" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.946745 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 10:18:17 crc kubenswrapper[4943]: W1204 10:18:17.966744 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58b0ae05_55dc_4238_9163_c8e06974ffca.slice/crio-14de82752c3f80ce09390306e7e94d76d85a7ddcb61bc69d195c2a5bfdc7c861 WatchSource:0}: Error finding container 14de82752c3f80ce09390306e7e94d76d85a7ddcb61bc69d195c2a5bfdc7c861: Status 404 returned error can't find the container with id 14de82752c3f80ce09390306e7e94d76d85a7ddcb61bc69d195c2a5bfdc7c861 Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.969047 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.987451 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 10:18:17 crc kubenswrapper[4943]: I1204 10:18:17.995504 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-serving-cert\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.006880 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.010267 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-config\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.034468 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.046070 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.046800 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.051478 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-service-ca-bundle\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.067175 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.086965 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.125442 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2g96c"] Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.125761 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d85wc\" (UniqueName: \"kubernetes.io/projected/1f2720fe-737e-4562-9fc1-d0e52e0d99a2-kube-api-access-d85wc\") pod \"machine-api-operator-5694c8668f-8sdsb\" (UID: \"1f2720fe-737e-4562-9fc1-d0e52e0d99a2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:18 crc kubenswrapper[4943]: W1204 10:18:18.134724 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8490c962_38c0_4be5_b940_d4c4d08c8a55.slice/crio-ef05790ae575ae735aad63f68e6efaae58dd78ad4799e8abbc7e3ebe1915b9b1 WatchSource:0}: Error finding container ef05790ae575ae735aad63f68e6efaae58dd78ad4799e8abbc7e3ebe1915b9b1: Status 404 returned error can't find the container with id ef05790ae575ae735aad63f68e6efaae58dd78ad4799e8abbc7e3ebe1915b9b1 Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.137640 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.151742 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.167079 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.186979 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.205699 4943 request.go:700] Waited for 1.006444304s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca/configmaps?fieldSelector=metadata.name%3Dsigning-cabundle&limit=500&resourceVersion=0 Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.207620 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.212859 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/db1b9fc6-529d-4a5c-9ef9-b91b6f872225-signing-cabundle\") pod \"service-ca-9c57cc56f-bnxzq\" (UID: \"db1b9fc6-529d-4a5c-9ef9-b91b6f872225\") " pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.227329 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.252283 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.253979 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.266774 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.287041 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.296590 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/db1b9fc6-529d-4a5c-9ef9-b91b6f872225-signing-key\") pod \"service-ca-9c57cc56f-bnxzq\" (UID: \"db1b9fc6-529d-4a5c-9ef9-b91b6f872225\") " pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.306674 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.327326 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.347606 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.366536 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.381803 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e5550ab-55c3-4147-871f-3e49bf38ee55-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-hjcd2\" (UID: \"6e5550ab-55c3-4147-871f-3e49bf38ee55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.383726 4943 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.383799 4943 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.383867 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e188d23e-750e-4c02-9388-0c8bc4a2906d-srv-cert podName:e188d23e-750e-4c02-9388-0c8bc4a2906d nodeName:}" failed. No retries permitted until 2025-12-04 10:18:18.883828858 +0000 UTC m=+147.472804726 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/e188d23e-750e-4c02-9388-0c8bc4a2906d-srv-cert") pod "catalog-operator-68c6474976-fzpv5" (UID: "e188d23e-750e-4c02-9388-0c8bc4a2906d") : failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.383744 4943 configmap.go:193] Couldn't get configMap openshift-etcd-operator/etcd-service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.383899 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-config-volume podName:a3b7dc03-7238-49ce-8ad6-1eabd0198a87 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:18.8838892 +0000 UTC m=+147.472865178 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-config-volume") pod "collect-profiles-29414055-z8m4l" (UID: "a3b7dc03-7238-49ce-8ad6-1eabd0198a87") : failed to sync configmap cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.383944 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-service-ca podName:7874756d-1da1-481a-be6e-df6034c4cac5 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:18.883914641 +0000 UTC m=+147.472890569 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-service-ca" (UniqueName: "kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-service-ca") pod "etcd-operator-b45778765-rbhdx" (UID: "7874756d-1da1-481a-be6e-df6034c4cac5") : failed to sync configmap cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.384996 4943 configmap.go:193] Couldn't get configMap openshift-etcd-operator/etcd-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.385070 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-config podName:7874756d-1da1-481a-be6e-df6034c4cac5 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:18.885050999 +0000 UTC m=+147.474026937 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-config") pod "etcd-operator-b45778765-rbhdx" (UID: "7874756d-1da1-481a-be6e-df6034c4cac5") : failed to sync configmap cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.385275 4943 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.385413 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-apiservice-cert podName:bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:18.885388397 +0000 UTC m=+147.474364295 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-apiservice-cert") pod "packageserver-d55dfcdfc-dmb24" (UID: "bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021") : failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.385468 4943 secret.go:188] Couldn't get secret openshift-machine-config-operator/mcc-proxy-tls: failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.385512 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0c0b640d-f71b-4c0e-8b92-e592722ccee5-proxy-tls podName:0c0b640d-f71b-4c0e-8b92-e592722ccee5 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:18.88549937 +0000 UTC m=+147.474475278 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/0c0b640d-f71b-4c0e-8b92-e592722ccee5-proxy-tls") pod "machine-config-controller-84d6567774-kvmzb" (UID: "0c0b640d-f71b-4c0e-8b92-e592722ccee5") : failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.385895 4943 secret.go:188] Couldn't get secret openshift-etcd-operator/etcd-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.385935 4943 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.385942 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7874756d-1da1-481a-be6e-df6034c4cac5-serving-cert podName:7874756d-1da1-481a-be6e-df6034c4cac5 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:18.88593196 +0000 UTC m=+147.474907878 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/7874756d-1da1-481a-be6e-df6034c4cac5-serving-cert") pod "etcd-operator-b45778765-rbhdx" (UID: "7874756d-1da1-481a-be6e-df6034c4cac5") : failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.386039 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-webhook-cert podName:bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:18.886014582 +0000 UTC m=+147.474990520 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-webhook-cert") pod "packageserver-d55dfcdfc-dmb24" (UID: "bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021") : failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.386666 4943 configmap.go:193] Couldn't get configMap openshift-etcd-operator/etcd-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.386814 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-ca podName:7874756d-1da1-481a-be6e-df6034c4cac5 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:18.886800581 +0000 UTC m=+147.475776509 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-ca" (UniqueName: "kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-ca") pod "etcd-operator-b45778765-rbhdx" (UID: "7874756d-1da1-481a-be6e-df6034c4cac5") : failed to sync configmap cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.386748 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.389181 4943 secret.go:188] Couldn't get secret openshift-ingress-operator/metrics-tls: failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.389176 4943 secret.go:188] Couldn't get secret openshift-etcd-operator/etcd-client: failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.389340 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-metrics-tls podName:149c7312-7be7-47d0-a2a4-bd0b0f588e3a nodeName:}" failed. No retries permitted until 2025-12-04 10:18:18.889327073 +0000 UTC m=+147.478302991 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-metrics-tls") pod "ingress-operator-5b745b69d9-hkrzs" (UID: "149c7312-7be7-47d0-a2a4-bd0b0f588e3a") : failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: E1204 10:18:18.389491 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-client podName:7874756d-1da1-481a-be6e-df6034c4cac5 nodeName:}" failed. No retries permitted until 2025-12-04 10:18:18.889468467 +0000 UTC m=+147.478444375 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-client") pod "etcd-operator-b45778765-rbhdx" (UID: "7874756d-1da1-481a-be6e-df6034c4cac5") : failed to sync secret cache: timed out waiting for the condition Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.391483 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e5550ab-55c3-4147-871f-3e49bf38ee55-config\") pod \"kube-controller-manager-operator-78b949d7b-hjcd2\" (UID: \"6e5550ab-55c3-4147-871f-3e49bf38ee55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.406479 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.426510 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.447474 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.467464 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.487653 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.507359 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.527475 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.547857 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.567882 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.587441 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.608092 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.627593 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.629068 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.629183 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.647803 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.667774 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.687512 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.708252 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.724526 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" event={"ID":"58b0ae05-55dc-4238-9163-c8e06974ffca","Type":"ContainerStarted","Data":"14de82752c3f80ce09390306e7e94d76d85a7ddcb61bc69d195c2a5bfdc7c861"} Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.725340 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" event={"ID":"fd459946-dd06-42f3-896c-8ae5b7394b33","Type":"ContainerStarted","Data":"98269efd2678af80f7b3942a937a01a506044348ccb6ed7161bd46fc93a7109f"} Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.726189 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" event={"ID":"b66208b0-ad95-479b-8fae-c239c7d91fdb","Type":"ContainerStarted","Data":"df3f5d8578de6225871a014de664631f65035bdd690296bb9229fe1c5cf8447f"} Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.726892 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" event={"ID":"8490c962-38c0-4be5-b940-d4c4d08c8a55","Type":"ContainerStarted","Data":"ef05790ae575ae735aad63f68e6efaae58dd78ad4799e8abbc7e3ebe1915b9b1"} Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.727338 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.746988 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.767807 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.787735 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.807605 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.826469 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.847470 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.854281 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-trusted-ca\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.866648 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.887603 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.907774 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.913238 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-webhook-cert\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.913585 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-config-volume\") pod \"collect-profiles-29414055-z8m4l\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.914729 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0c0b640d-f71b-4c0e-8b92-e592722ccee5-proxy-tls\") pod \"machine-config-controller-84d6567774-kvmzb\" (UID: \"0c0b640d-f71b-4c0e-8b92-e592722ccee5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.914633 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-config-volume\") pod \"collect-profiles-29414055-z8m4l\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.918095 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-ca\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.918883 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-ca\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.919149 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-client\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.919669 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-webhook-cert\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.919966 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-metrics-tls\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.920237 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e188d23e-750e-4c02-9388-0c8bc4a2906d-srv-cert\") pod \"catalog-operator-68c6474976-fzpv5\" (UID: \"e188d23e-750e-4c02-9388-0c8bc4a2906d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.920295 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7874756d-1da1-481a-be6e-df6034c4cac5-serving-cert\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.920384 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-config\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.920462 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-apiservice-cert\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.920499 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-service-ca\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.921467 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-service-ca\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.921806 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0c0b640d-f71b-4c0e-8b92-e592722ccee5-proxy-tls\") pod \"machine-config-controller-84d6567774-kvmzb\" (UID: \"0c0b640d-f71b-4c0e-8b92-e592722ccee5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.922418 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7874756d-1da1-481a-be6e-df6034c4cac5-config\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.923334 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7874756d-1da1-481a-be6e-df6034c4cac5-etcd-client\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.923492 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-metrics-tls\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.924115 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e188d23e-750e-4c02-9388-0c8bc4a2906d-srv-cert\") pod \"catalog-operator-68c6474976-fzpv5\" (UID: \"e188d23e-750e-4c02-9388-0c8bc4a2906d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.925305 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-apiservice-cert\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.926337 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7874756d-1da1-481a-be6e-df6034c4cac5-serving-cert\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.926825 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.947011 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 10:18:18 crc kubenswrapper[4943]: I1204 10:18:18.987790 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.007938 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.028178 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-8sdsb"] Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.028380 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.047234 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.066522 4943 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.087603 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.118831 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.127121 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.147409 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.172274 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.205817 4943 request.go:700] Waited for 1.822747097s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/serviceaccounts/collect-profiles/token Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.222954 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nfln\" (UniqueName: \"kubernetes.io/projected/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-kube-api-access-7nfln\") pod \"collect-profiles-29414055-z8m4l\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.240222 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cdf2c613-8426-4c4e-8cfe-bc015e17e47e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-x5dsb\" (UID: \"cdf2c613-8426-4c4e-8cfe-bc015e17e47e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.265937 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxbwf\" (UniqueName: \"kubernetes.io/projected/89aaae4b-99da-4820-b024-a013c629c63f-kube-api-access-fxbwf\") pod \"olm-operator-6b444d44fb-pxxxc\" (UID: \"89aaae4b-99da-4820-b024-a013c629c63f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.285423 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5wbr\" (UniqueName: \"kubernetes.io/projected/0c0b640d-f71b-4c0e-8b92-e592722ccee5-kube-api-access-n5wbr\") pod \"machine-config-controller-84d6567774-kvmzb\" (UID: \"0c0b640d-f71b-4c0e-8b92-e592722ccee5\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.303600 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pnm9\" (UniqueName: \"kubernetes.io/projected/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-kube-api-access-8pnm9\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.309626 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.324840 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.329820 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz568\" (UniqueName: \"kubernetes.io/projected/11c1150e-cb65-4c9d-98b6-6c36f5092307-kube-api-access-bz568\") pod \"kube-storage-version-migrator-operator-b67b599dd-lnrw7\" (UID: \"11c1150e-cb65-4c9d-98b6-6c36f5092307\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.343230 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq479\" (UniqueName: \"kubernetes.io/projected/3e8ecae4-2aee-4125-a16c-d4f340d840de-kube-api-access-qq479\") pod \"package-server-manager-789f6589d5-ws8hv\" (UID: \"3e8ecae4-2aee-4125-a16c-d4f340d840de\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.363793 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b3c8ac74-bfd7-494b-9181-36832b3ffa98-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-779v8\" (UID: \"b3c8ac74-bfd7-494b-9181-36832b3ffa98\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.388338 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5mtt\" (UniqueName: \"kubernetes.io/projected/7ec6444d-9ad0-4310-974a-d8525e16e950-kube-api-access-l5mtt\") pod \"machine-approver-56656f9798-h9rws\" (UID: \"7ec6444d-9ad0-4310-974a-d8525e16e950\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.401909 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfztz\" (UniqueName: \"kubernetes.io/projected/bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021-kube-api-access-xfztz\") pod \"packageserver-d55dfcdfc-dmb24\" (UID: \"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.426518 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2s7d\" (UniqueName: \"kubernetes.io/projected/14bd93bd-8f38-48c3-9508-6c1ab9e65628-kube-api-access-g2s7d\") pod \"oauth-openshift-558db77b4-wkldv\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.445108 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9rrn\" (UniqueName: \"kubernetes.io/projected/7019146d-e522-4815-9642-9c9667aba409-kube-api-access-g9rrn\") pod \"dns-operator-744455d44c-ml4z6\" (UID: \"7019146d-e522-4815-9642-9c9667aba409\") " pod="openshift-dns-operator/dns-operator-744455d44c-ml4z6" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.451832 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.451840 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.458791 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.466593 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.473767 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx5tf\" (UniqueName: \"kubernetes.io/projected/db1b9fc6-529d-4a5c-9ef9-b91b6f872225-kube-api-access-qx5tf\") pod \"service-ca-9c57cc56f-bnxzq\" (UID: \"db1b9fc6-529d-4a5c-9ef9-b91b6f872225\") " pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.478773 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.487927 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br2tk\" (UniqueName: \"kubernetes.io/projected/3d93745e-ade6-45eb-b5e3-9b4270483319-kube-api-access-br2tk\") pod \"apiserver-76f77b778f-5bgm4\" (UID: \"3d93745e-ade6-45eb-b5e3-9b4270483319\") " pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.505003 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/149c7312-7be7-47d0-a2a4-bd0b0f588e3a-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hkrzs\" (UID: \"149c7312-7be7-47d0-a2a4-bd0b0f588e3a\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.550419 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsmwg\" (UniqueName: \"kubernetes.io/projected/d93e50fc-a2f7-4ed5-ad71-7590b393c71f-kube-api-access-qsmwg\") pod \"openshift-config-operator-7777fb866f-89wdj\" (UID: \"d93e50fc-a2f7-4ed5-ad71-7590b393c71f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.563730 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llrpq\" (UniqueName: \"kubernetes.io/projected/b97e1699-6f5e-461d-b3e0-4afaef630174-kube-api-access-llrpq\") pod \"openshift-apiserver-operator-796bbdcf4f-cm2tk\" (UID: \"b97e1699-6f5e-461d-b3e0-4afaef630174\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.575320 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.577659 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cefa13a6-8806-4b35-9083-df5aa7d45f56-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-snz6f\" (UID: \"cefa13a6-8806-4b35-9083-df5aa7d45f56\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.592285 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.592529 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp5tr\" (UniqueName: \"kubernetes.io/projected/7874756d-1da1-481a-be6e-df6034c4cac5-kube-api-access-xp5tr\") pod \"etcd-operator-b45778765-rbhdx\" (UID: \"7874756d-1da1-481a-be6e-df6034c4cac5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.600488 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.616309 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4gk2\" (UniqueName: \"kubernetes.io/projected/f93fafb8-e1d5-4ef2-a443-b3cfe19b2bda-kube-api-access-f4gk2\") pod \"migrator-59844c95c7-28rrg\" (UID: \"f93fafb8-e1d5-4ef2-a443-b3cfe19b2bda\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-28rrg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.667351 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4f4s\" (UniqueName: \"kubernetes.io/projected/b3c8ac74-bfd7-494b-9181-36832b3ffa98-kube-api-access-j4f4s\") pod \"cluster-image-registry-operator-dc59b4c8b-779v8\" (UID: \"b3c8ac74-bfd7-494b-9181-36832b3ffa98\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.669643 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.670511 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.670775 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.671346 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-28rrg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.672606 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.705884 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ml4z6" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.706433 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.713954 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szwh5\" (UniqueName: \"kubernetes.io/projected/560d2fc0-5e01-4d10-9499-72144de110f6-kube-api-access-szwh5\") pod \"machine-config-operator-74547568cd-fqmc8\" (UID: \"560d2fc0-5e01-4d10-9499-72144de110f6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.715360 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc5ss\" (UniqueName: \"kubernetes.io/projected/6c9f637d-2d43-4bb6-9db2-1831bd8e7573-kube-api-access-zc5ss\") pod \"console-operator-58897d9998-tf5g6\" (UID: \"6c9f637d-2d43-4bb6-9db2-1831bd8e7573\") " pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.717449 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77jvx\" (UniqueName: \"kubernetes.io/projected/1abfb759-04f9-4885-b150-82355f8cf866-kube-api-access-77jvx\") pod \"downloads-7954f5f757-grw5p\" (UID: \"1abfb759-04f9-4885-b150-82355f8cf866\") " pod="openshift-console/downloads-7954f5f757-grw5p" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.732058 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.734127 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtjkn\" (UniqueName: \"kubernetes.io/projected/7848703f-6d8c-4915-8f42-e6222f2a889a-kube-api-access-xtjkn\") pod \"console-f9d7485db-kdt7l\" (UID: \"7848703f-6d8c-4915-8f42-e6222f2a889a\") " pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.734475 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.741630 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.777062 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l"] Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.807955 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e5550ab-55c3-4147-871f-3e49bf38ee55-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-hjcd2\" (UID: \"6e5550ab-55c3-4147-871f-3e49bf38ee55\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.814761 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.815339 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wgml\" (UniqueName: \"kubernetes.io/projected/c0352a66-8ffd-4949-ba0f-d8e6cb933ee4-kube-api-access-8wgml\") pod \"service-ca-operator-777779d784-hgd88\" (UID: \"c0352a66-8ffd-4949-ba0f-d8e6cb933ee4\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.815651 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz5jv\" (UniqueName: \"kubernetes.io/projected/0e7f7c48-3d22-4aa4-8d06-f265758b3d36-kube-api-access-kz5jv\") pod \"authentication-operator-69f744f599-8nftj\" (UID: \"0e7f7c48-3d22-4aa4-8d06-f265758b3d36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.815677 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dn2r\" (UniqueName: \"kubernetes.io/projected/5ef5c158-a074-47df-a4bb-2e1cab760745-kube-api-access-8dn2r\") pod \"control-plane-machine-set-operator-78cbb6b69f-vhxkd\" (UID: \"5ef5c158-a074-47df-a4bb-2e1cab760745\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.825821 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" event={"ID":"8490c962-38c0-4be5-b940-d4c4d08c8a55","Type":"ContainerStarted","Data":"6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a"} Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.826542 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.828806 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c7m9\" (UniqueName: \"kubernetes.io/projected/931858b0-187e-441e-a961-80c7cf44d916-kube-api-access-8c7m9\") pod \"cluster-samples-operator-665b6dd947-cx7l7\" (UID: \"931858b0-187e-441e-a961-80c7cf44d916\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.832286 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" event={"ID":"58b0ae05-55dc-4238-9163-c8e06974ffca","Type":"ContainerStarted","Data":"df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36"} Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.833143 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjk27\" (UniqueName: \"kubernetes.io/projected/7f0e5e19-36d5-4553-896d-c5057c8fe1c3-kube-api-access-zjk27\") pod \"multus-admission-controller-857f4d67dd-m4262\" (UID: \"7f0e5e19-36d5-4553-896d-c5057c8fe1c3\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-m4262" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.833524 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.854450 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-m4262" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.863854 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" event={"ID":"fd459946-dd06-42f3-896c-8ae5b7394b33","Type":"ContainerStarted","Data":"8ead31ca2d48b40cb9857ef38a2fa0c66cbefa5762ef05f971166a54bb8dcb20"} Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.865438 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhd9q\" (UniqueName: \"kubernetes.io/projected/e188d23e-750e-4c02-9388-0c8bc4a2906d-kube-api-access-rhd9q\") pod \"catalog-operator-68c6474976-fzpv5\" (UID: \"e188d23e-750e-4c02-9388-0c8bc4a2906d\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.866788 4943 generic.go:334] "Generic (PLEG): container finished" podID="b66208b0-ad95-479b-8fae-c239c7d91fdb" containerID="068eb4931abc68f74df26d16bb16f21f11faafb3bbc55102f514d9c57978f79b" exitCode=0 Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.866964 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.867796 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" event={"ID":"b66208b0-ad95-479b-8fae-c239c7d91fdb","Type":"ContainerDied","Data":"068eb4931abc68f74df26d16bb16f21f11faafb3bbc55102f514d9c57978f79b"} Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.870006 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-grw5p" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.878753 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1a3d277-1dd7-4970-8017-4dca49e1b358-config-volume\") pod \"dns-default-fs4kj\" (UID: \"d1a3d277-1dd7-4970-8017-4dca49e1b358\") " pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.878791 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6da7528-9b01-40b4-88d3-e78a3c178300-trusted-ca\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.878817 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46dfg\" (UniqueName: \"kubernetes.io/projected/d1a3d277-1dd7-4970-8017-4dca49e1b358-kube-api-access-46dfg\") pod \"dns-default-fs4kj\" (UID: \"d1a3d277-1dd7-4970-8017-4dca49e1b358\") " pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.878866 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.878894 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a6da7528-9b01-40b4-88d3-e78a3c178300-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.878937 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4v8t\" (UniqueName: \"kubernetes.io/projected/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-kube-api-access-r4v8t\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.878997 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-stats-auth\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.879036 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3e183218-08a1-4890-a6af-d56f3c7db2be-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ph88v\" (UID: \"3e183218-08a1-4890-a6af-d56f3c7db2be\") " pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.879091 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-registry-tls\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.879112 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a6da7528-9b01-40b4-88d3-e78a3c178300-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.879164 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-metrics-certs\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.879187 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-service-ca-bundle\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.879232 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mvp7\" (UniqueName: \"kubernetes.io/projected/3e183218-08a1-4890-a6af-d56f3c7db2be-kube-api-access-8mvp7\") pod \"marketplace-operator-79b997595-ph88v\" (UID: \"3e183218-08a1-4890-a6af-d56f3c7db2be\") " pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.879274 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-bound-sa-token\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.879296 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-default-certificate\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.879333 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d1a3d277-1dd7-4970-8017-4dca49e1b358-metrics-tls\") pod \"dns-default-fs4kj\" (UID: \"d1a3d277-1dd7-4970-8017-4dca49e1b358\") " pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.879362 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxxrv\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-kube-api-access-sxxrv\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.879443 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e183218-08a1-4890-a6af-d56f3c7db2be-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ph88v\" (UID: \"3e183218-08a1-4890-a6af-d56f3c7db2be\") " pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.879580 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a6da7528-9b01-40b4-88d3-e78a3c178300-registry-certificates\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: E1204 10:18:19.880313 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:20.38029812 +0000 UTC m=+148.969273988 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.882863 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" event={"ID":"1f2720fe-737e-4562-9fc1-d0e52e0d99a2","Type":"ContainerStarted","Data":"bdc43aeedc45d41295c25a1827162dfd7b09b6edb6ef13153d83a0bb51167b5e"} Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.882911 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" event={"ID":"1f2720fe-737e-4562-9fc1-d0e52e0d99a2","Type":"ContainerStarted","Data":"f9c59d0c9a60071df4d5210e12ef67a1479ed30a90f49a02583efb3f8d7f911e"} Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.882936 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" event={"ID":"1f2720fe-737e-4562-9fc1-d0e52e0d99a2","Type":"ContainerStarted","Data":"7bc542aa035031e92769dfbcea99742842475daf42394d20500b8fbd8f0e3599"} Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.883600 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.887942 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.927035 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.934471 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.946272 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.957880 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb"] Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.979459 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.979974 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.980752 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:19 crc kubenswrapper[4943]: E1204 10:18:19.980848 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:20.480771702 +0000 UTC m=+149.069747570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.981013 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e183218-08a1-4890-a6af-d56f3c7db2be-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ph88v\" (UID: \"3e183218-08a1-4890-a6af-d56f3c7db2be\") " pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.981082 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-mountpoint-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.981106 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlr5r\" (UniqueName: \"kubernetes.io/projected/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-kube-api-access-rlr5r\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.981428 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a6da7528-9b01-40b4-88d3-e78a3c178300-registry-certificates\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.981451 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-plugins-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.981499 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-socket-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.981847 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1a3d277-1dd7-4970-8017-4dca49e1b358-config-volume\") pod \"dns-default-fs4kj\" (UID: \"d1a3d277-1dd7-4970-8017-4dca49e1b358\") " pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.981914 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6da7528-9b01-40b4-88d3-e78a3c178300-trusted-ca\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.981934 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46dfg\" (UniqueName: \"kubernetes.io/projected/d1a3d277-1dd7-4970-8017-4dca49e1b358-kube-api-access-46dfg\") pod \"dns-default-fs4kj\" (UID: \"d1a3d277-1dd7-4970-8017-4dca49e1b358\") " pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.981953 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d64e4f2a-d704-4e63-992f-a6fe10c3ca87-certs\") pod \"machine-config-server-xrlzc\" (UID: \"d64e4f2a-d704-4e63-992f-a6fe10c3ca87\") " pod="openshift-machine-config-operator/machine-config-server-xrlzc" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.982644 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e183218-08a1-4890-a6af-d56f3c7db2be-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ph88v\" (UID: \"3e183218-08a1-4890-a6af-d56f3c7db2be\") " pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.984452 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a6da7528-9b01-40b4-88d3-e78a3c178300-registry-certificates\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.985028 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1a3d277-1dd7-4970-8017-4dca49e1b358-config-volume\") pod \"dns-default-fs4kj\" (UID: \"d1a3d277-1dd7-4970-8017-4dca49e1b358\") " pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.988338 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.988480 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a6da7528-9b01-40b4-88d3-e78a3c178300-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.988741 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-csi-data-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.989152 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a6da7528-9b01-40b4-88d3-e78a3c178300-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: E1204 10:18:19.989335 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:20.489312181 +0000 UTC m=+149.078288109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.989370 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4v8t\" (UniqueName: \"kubernetes.io/projected/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-kube-api-access-r4v8t\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.990080 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-registration-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.991372 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-stats-auth\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.996112 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3e183218-08a1-4890-a6af-d56f3c7db2be-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ph88v\" (UID: \"3e183218-08a1-4890-a6af-d56f3c7db2be\") " pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.997168 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-registry-tls\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.997461 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d64e4f2a-d704-4e63-992f-a6fe10c3ca87-node-bootstrap-token\") pod \"machine-config-server-xrlzc\" (UID: \"d64e4f2a-d704-4e63-992f-a6fe10c3ca87\") " pod="openshift-machine-config-operator/machine-config-server-xrlzc" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.997558 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a6da7528-9b01-40b4-88d3-e78a3c178300-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.997850 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrv6t\" (UniqueName: \"kubernetes.io/projected/f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2-kube-api-access-qrv6t\") pod \"ingress-canary-6jx2n\" (UID: \"f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2\") " pod="openshift-ingress-canary/ingress-canary-6jx2n" Dec 04 10:18:19 crc kubenswrapper[4943]: I1204 10:18:19.998065 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-metrics-certs\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.003291 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-service-ca-bundle\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.004578 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mvp7\" (UniqueName: \"kubernetes.io/projected/3e183218-08a1-4890-a6af-d56f3c7db2be-kube-api-access-8mvp7\") pod \"marketplace-operator-79b997595-ph88v\" (UID: \"3e183218-08a1-4890-a6af-d56f3c7db2be\") " pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.008678 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-stats-auth\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.009069 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-default-certificate\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.009099 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-bound-sa-token\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.009143 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d1a3d277-1dd7-4970-8017-4dca49e1b358-metrics-tls\") pod \"dns-default-fs4kj\" (UID: \"d1a3d277-1dd7-4970-8017-4dca49e1b358\") " pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.009342 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnj46\" (UniqueName: \"kubernetes.io/projected/d64e4f2a-d704-4e63-992f-a6fe10c3ca87-kube-api-access-cnj46\") pod \"machine-config-server-xrlzc\" (UID: \"d64e4f2a-d704-4e63-992f-a6fe10c3ca87\") " pod="openshift-machine-config-operator/machine-config-server-xrlzc" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.009571 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxxrv\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-kube-api-access-sxxrv\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.009754 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2-cert\") pod \"ingress-canary-6jx2n\" (UID: \"f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2\") " pod="openshift-ingress-canary/ingress-canary-6jx2n" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.011589 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6da7528-9b01-40b4-88d3-e78a3c178300-trusted-ca\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.016710 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a6da7528-9b01-40b4-88d3-e78a3c178300-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.018275 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.020230 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.025472 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-service-ca-bundle\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.026277 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-registry-tls\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.035288 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-default-certificate\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.051088 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d1a3d277-1dd7-4970-8017-4dca49e1b358-metrics-tls\") pod \"dns-default-fs4kj\" (UID: \"d1a3d277-1dd7-4970-8017-4dca49e1b358\") " pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.056354 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7"] Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.069973 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3e183218-08a1-4890-a6af-d56f3c7db2be-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ph88v\" (UID: \"3e183218-08a1-4890-a6af-d56f3c7db2be\") " pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.075398 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-bound-sa-token\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.078322 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4v8t\" (UniqueName: \"kubernetes.io/projected/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-kube-api-access-r4v8t\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.080312 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cb5f8eaa-9da1-4a88-b277-b8f76e86123d-metrics-certs\") pod \"router-default-5444994796-zkjfk\" (UID: \"cb5f8eaa-9da1-4a88-b277-b8f76e86123d\") " pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.094817 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46dfg\" (UniqueName: \"kubernetes.io/projected/d1a3d277-1dd7-4970-8017-4dca49e1b358-kube-api-access-46dfg\") pod \"dns-default-fs4kj\" (UID: \"d1a3d277-1dd7-4970-8017-4dca49e1b358\") " pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.131722 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.131966 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-registration-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.132005 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d64e4f2a-d704-4e63-992f-a6fe10c3ca87-node-bootstrap-token\") pod \"machine-config-server-xrlzc\" (UID: \"d64e4f2a-d704-4e63-992f-a6fe10c3ca87\") " pod="openshift-machine-config-operator/machine-config-server-xrlzc" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.132033 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrv6t\" (UniqueName: \"kubernetes.io/projected/f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2-kube-api-access-qrv6t\") pod \"ingress-canary-6jx2n\" (UID: \"f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2\") " pod="openshift-ingress-canary/ingress-canary-6jx2n" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.132089 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnj46\" (UniqueName: \"kubernetes.io/projected/d64e4f2a-d704-4e63-992f-a6fe10c3ca87-kube-api-access-cnj46\") pod \"machine-config-server-xrlzc\" (UID: \"d64e4f2a-d704-4e63-992f-a6fe10c3ca87\") " pod="openshift-machine-config-operator/machine-config-server-xrlzc" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.132109 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2-cert\") pod \"ingress-canary-6jx2n\" (UID: \"f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2\") " pod="openshift-ingress-canary/ingress-canary-6jx2n" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.132141 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-mountpoint-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.132160 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlr5r\" (UniqueName: \"kubernetes.io/projected/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-kube-api-access-rlr5r\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.132221 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-plugins-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.132238 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-socket-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.132262 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d64e4f2a-d704-4e63-992f-a6fe10c3ca87-certs\") pod \"machine-config-server-xrlzc\" (UID: \"d64e4f2a-d704-4e63-992f-a6fe10c3ca87\") " pod="openshift-machine-config-operator/machine-config-server-xrlzc" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.132316 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-csi-data-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.132465 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-csi-data-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: E1204 10:18:20.132541 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:20.632524202 +0000 UTC m=+149.221500070 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.132689 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-registration-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.134610 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-mountpoint-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.134803 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-plugins-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.134914 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-socket-dir\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.143634 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2-cert\") pod \"ingress-canary-6jx2n\" (UID: \"f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2\") " pod="openshift-ingress-canary/ingress-canary-6jx2n" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.145149 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d64e4f2a-d704-4e63-992f-a6fe10c3ca87-node-bootstrap-token\") pod \"machine-config-server-xrlzc\" (UID: \"d64e4f2a-d704-4e63-992f-a6fe10c3ca87\") " pod="openshift-machine-config-operator/machine-config-server-xrlzc" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.146263 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d64e4f2a-d704-4e63-992f-a6fe10c3ca87-certs\") pod \"machine-config-server-xrlzc\" (UID: \"d64e4f2a-d704-4e63-992f-a6fe10c3ca87\") " pod="openshift-machine-config-operator/machine-config-server-xrlzc" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.155837 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxxrv\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-kube-api-access-sxxrv\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.175461 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mvp7\" (UniqueName: \"kubernetes.io/projected/3e183218-08a1-4890-a6af-d56f3c7db2be-kube-api-access-8mvp7\") pod \"marketplace-operator-79b997595-ph88v\" (UID: \"3e183218-08a1-4890-a6af-d56f3c7db2be\") " pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:20 crc kubenswrapper[4943]: W1204 10:18:20.191157 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c0b640d_f71b_4c0e_8b92_e592722ccee5.slice/crio-633026da9217d1697bf412e3cad8cc2fa86989e1a93930b6759b8ffdcae590f2 WatchSource:0}: Error finding container 633026da9217d1697bf412e3cad8cc2fa86989e1a93930b6759b8ffdcae590f2: Status 404 returned error can't find the container with id 633026da9217d1697bf412e3cad8cc2fa86989e1a93930b6759b8ffdcae590f2 Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.218491 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnj46\" (UniqueName: \"kubernetes.io/projected/d64e4f2a-d704-4e63-992f-a6fe10c3ca87-kube-api-access-cnj46\") pod \"machine-config-server-xrlzc\" (UID: \"d64e4f2a-d704-4e63-992f-a6fe10c3ca87\") " pod="openshift-machine-config-operator/machine-config-server-xrlzc" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.233886 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:20 crc kubenswrapper[4943]: E1204 10:18:20.234351 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:20.734337936 +0000 UTC m=+149.323313794 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.239484 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.255089 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.278513 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrv6t\" (UniqueName: \"kubernetes.io/projected/f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2-kube-api-access-qrv6t\") pod \"ingress-canary-6jx2n\" (UID: \"f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2\") " pod="openshift-ingress-canary/ingress-canary-6jx2n" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.286240 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-6jx2n" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.299549 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlr5r\" (UniqueName: \"kubernetes.io/projected/d9c45054-0f0f-4bc4-9efd-5b45c98dda8b-kube-api-access-rlr5r\") pod \"csi-hostpathplugin-rqdpv\" (UID: \"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b\") " pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.299784 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xrlzc" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.304768 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc"] Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.337375 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv"] Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.339676 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:20 crc kubenswrapper[4943]: E1204 10:18:20.339889 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:20.839857813 +0000 UTC m=+149.428833681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.340026 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:20 crc kubenswrapper[4943]: E1204 10:18:20.340410 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:20.840395896 +0000 UTC m=+149.429371764 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.442805 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:20 crc kubenswrapper[4943]: E1204 10:18:20.443304 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:20.943284146 +0000 UTC m=+149.532260014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.443424 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.545414 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:20 crc kubenswrapper[4943]: E1204 10:18:20.545847 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.045822479 +0000 UTC m=+149.634798547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.590433 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.647038 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:20 crc kubenswrapper[4943]: E1204 10:18:20.647611 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.147593443 +0000 UTC m=+149.736569311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.751249 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:20 crc kubenswrapper[4943]: E1204 10:18:20.751922 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.25190587 +0000 UTC m=+149.840881738 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.778810 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb"] Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.784693 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs"] Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.852737 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:20 crc kubenswrapper[4943]: E1204 10:18:20.852913 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.352875914 +0000 UTC m=+149.941851782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.853243 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:20 crc kubenswrapper[4943]: E1204 10:18:20.853633 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.353617442 +0000 UTC m=+149.942593310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.924393 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" event={"ID":"7ec6444d-9ad0-4310-974a-d8525e16e950","Type":"ContainerStarted","Data":"e1f9ca7fbee4309c7a82a5928d2dc5cde1e22679850cb384adda843eb3d978d1"} Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.937352 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" event={"ID":"0c0b640d-f71b-4c0e-8b92-e592722ccee5","Type":"ContainerStarted","Data":"e5aeb2f901a9701d1b4391fb25b78efc56749c89a516e0c946e586bc5560e7d3"} Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.937410 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" event={"ID":"0c0b640d-f71b-4c0e-8b92-e592722ccee5","Type":"ContainerStarted","Data":"633026da9217d1697bf412e3cad8cc2fa86989e1a93930b6759b8ffdcae590f2"} Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.961871 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.961869 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xrlzc" event={"ID":"d64e4f2a-d704-4e63-992f-a6fe10c3ca87","Type":"ContainerStarted","Data":"9390cabe698af6d006717411783a63166cbe4d691bd7f287f884fcbf434afc1b"} Dec 04 10:18:20 crc kubenswrapper[4943]: E1204 10:18:20.962365 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.462345036 +0000 UTC m=+150.051320904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:20 crc kubenswrapper[4943]: I1204 10:18:20.980167 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" event={"ID":"89aaae4b-99da-4820-b024-a013c629c63f","Type":"ContainerStarted","Data":"1fdbf98f919e4172a655b99e8ca6ae73d64675557cbf89890c0d987e91e16a74"} Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.015060 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" event={"ID":"3e8ecae4-2aee-4125-a16c-d4f340d840de","Type":"ContainerStarted","Data":"f9cc506513fc6c8d3cae89b3eb8c9826df8ba0c4471d487635ae1fb86ac0b3a4"} Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.020728 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" event={"ID":"a3b7dc03-7238-49ce-8ad6-1eabd0198a87","Type":"ContainerStarted","Data":"02e6fee0a87f0fa26073935f809b8b393c46a4c1abd28554f89117833a8d6e9d"} Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.024380 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" event={"ID":"a3b7dc03-7238-49ce-8ad6-1eabd0198a87","Type":"ContainerStarted","Data":"1760c0af72bde1fa103a90b058e355793f1c07d1f45e33f9efe4687bca957c2d"} Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.117785 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:21 crc kubenswrapper[4943]: E1204 10:18:21.118906 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.618878012 +0000 UTC m=+150.207853880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.129067 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" event={"ID":"11c1150e-cb65-4c9d-98b6-6c36f5092307","Type":"ContainerStarted","Data":"a56efba9aaf65521ba8eca550182e17ac16fa04f7ccfbdffcc252c842e0c7f3a"} Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.129114 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" event={"ID":"11c1150e-cb65-4c9d-98b6-6c36f5092307","Type":"ContainerStarted","Data":"b5669400f111dfe4fc51595abdd26a7d77d8b4ce437948ee0259162172dfed1b"} Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.154298 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zkjfk" event={"ID":"cb5f8eaa-9da1-4a88-b277-b8f76e86123d","Type":"ContainerStarted","Data":"7047ad10493bd8c6ceb459413ce1a062cce3040b99fa24eaf3e739d570012de8"} Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.222646 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:21 crc kubenswrapper[4943]: E1204 10:18:21.225708 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.725681089 +0000 UTC m=+150.314656977 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.325418 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:21 crc kubenswrapper[4943]: E1204 10:18:21.327724 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.827710169 +0000 UTC m=+150.416686037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.379504 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-28rrg"] Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.390713 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wkldv"] Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.428591 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.428797 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.428836 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.428875 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.435590 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:18:21 crc kubenswrapper[4943]: E1204 10:18:21.436734 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:21.93671368 +0000 UTC m=+150.525689548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.462737 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.462891 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:18:21 crc kubenswrapper[4943]: W1204 10:18:21.511280 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf93fafb8_e1d5_4ef2_a443_b3cfe19b2bda.slice/crio-52bd0e7fc39b96387dd9d41aebe9c8969cb58b492b339745b3092bac7f3efbee WatchSource:0}: Error finding container 52bd0e7fc39b96387dd9d41aebe9c8969cb58b492b339745b3092bac7f3efbee: Status 404 returned error can't find the container with id 52bd0e7fc39b96387dd9d41aebe9c8969cb58b492b339745b3092bac7f3efbee Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.532456 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.532563 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:18:21 crc kubenswrapper[4943]: E1204 10:18:21.537388 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:22.037372317 +0000 UTC m=+150.626348185 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.584314 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24"] Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.599951 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.614636 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.625478 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-tf5g6"] Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.627899 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.633416 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:21 crc kubenswrapper[4943]: E1204 10:18:21.633782 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:22.133765279 +0000 UTC m=+150.722741147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.634939 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-bnxzq"] Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.651279 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-rbhdx"] Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.699426 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8"] Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.706866 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-89wdj"] Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.740179 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:21 crc kubenswrapper[4943]: E1204 10:18:21.740644 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:22.240629128 +0000 UTC m=+150.829604986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.833950 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" podStartSLOduration=126.833918984 podStartE2EDuration="2m6.833918984s" podCreationTimestamp="2025-12-04 10:16:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:21.832680323 +0000 UTC m=+150.421656191" watchObservedRunningTime="2025-12-04 10:18:21.833918984 +0000 UTC m=+150.422894852" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.842638 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:21 crc kubenswrapper[4943]: E1204 10:18:21.843061 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:22.343043877 +0000 UTC m=+150.932019745 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.862330 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-8sdsb" podStartSLOduration=127.86231033 podStartE2EDuration="2m7.86231033s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:21.860322331 +0000 UTC m=+150.449298199" watchObservedRunningTime="2025-12-04 10:18:21.86231033 +0000 UTC m=+150.451286198" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.893517 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.922376 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" podStartSLOduration=127.92234093 podStartE2EDuration="2m7.92234093s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:21.901372677 +0000 UTC m=+150.490348545" watchObservedRunningTime="2025-12-04 10:18:21.92234093 +0000 UTC m=+150.511316798" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.948371 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:21 crc kubenswrapper[4943]: E1204 10:18:21.949066 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:22.449047555 +0000 UTC m=+151.038023423 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.981331 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" podStartSLOduration=127.981313175 podStartE2EDuration="2m7.981313175s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:21.949279641 +0000 UTC m=+150.538255509" watchObservedRunningTime="2025-12-04 10:18:21.981313175 +0000 UTC m=+150.570289033" Dec 04 10:18:21 crc kubenswrapper[4943]: I1204 10:18:21.985971 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lnrw7" podStartSLOduration=127.985948079 podStartE2EDuration="2m7.985948079s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:21.980284261 +0000 UTC m=+150.569260129" watchObservedRunningTime="2025-12-04 10:18:21.985948079 +0000 UTC m=+150.574923937" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.051412 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:22 crc kubenswrapper[4943]: E1204 10:18:22.051575 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:22.551548848 +0000 UTC m=+151.140524716 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.051687 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:22 crc kubenswrapper[4943]: E1204 10:18:22.052019 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:22.552009949 +0000 UTC m=+151.140985817 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.153145 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:22 crc kubenswrapper[4943]: E1204 10:18:22.153865 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:22.653823883 +0000 UTC m=+151.242799751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.219987 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zkjfk" event={"ID":"cb5f8eaa-9da1-4a88-b277-b8f76e86123d","Type":"ContainerStarted","Data":"ea3eb42e23c061491acd41f82eabd63a88ef8685654ab90646527cbc3f15a9b1"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.235550 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" event={"ID":"b66208b0-ad95-479b-8fae-c239c7d91fdb","Type":"ContainerStarted","Data":"0e67fbe311c51d70298da9fd0abed0ad03ecb4c0b87eee18671f272c0ed8461b"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.245306 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.256118 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:22 crc kubenswrapper[4943]: E1204 10:18:22.257444 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:22.757428152 +0000 UTC m=+151.346404030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.262713 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-zkjfk" podStartSLOduration=128.262693861 podStartE2EDuration="2m8.262693861s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:22.262063126 +0000 UTC m=+150.851039014" watchObservedRunningTime="2025-12-04 10:18:22.262693861 +0000 UTC m=+150.851669729" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.263071 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-jbf9m" podStartSLOduration=128.26306484 podStartE2EDuration="2m8.26306484s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:22.082856685 +0000 UTC m=+150.671832553" watchObservedRunningTime="2025-12-04 10:18:22.26306484 +0000 UTC m=+150.852040708" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.325539 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-tf5g6" event={"ID":"6c9f637d-2d43-4bb6-9db2-1831bd8e7573","Type":"ContainerStarted","Data":"894f20f5d090ab60482f6ba1a768364aafacb5d55d1dadcd351bb6d024b132e1"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.345087 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" event={"ID":"14bd93bd-8f38-48c3-9508-6c1ab9e65628","Type":"ContainerStarted","Data":"3a4b6bfeeae45122b65be01796496703f4ab683dbd5090b41401134d15c2ee35"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.353945 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" podStartSLOduration=127.353913376 podStartE2EDuration="2m7.353913376s" podCreationTimestamp="2025-12-04 10:16:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:22.347997641 +0000 UTC m=+150.936973509" watchObservedRunningTime="2025-12-04 10:18:22.353913376 +0000 UTC m=+150.942889254" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.358010 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:22 crc kubenswrapper[4943]: E1204 10:18:22.359323 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:22.859295838 +0000 UTC m=+151.448271766 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.373430 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" event={"ID":"b3c8ac74-bfd7-494b-9181-36832b3ffa98","Type":"ContainerStarted","Data":"ee08ee2f7c0879ccce92660e18398d1f4febbd6725fc3d735af42b7695594713"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.397453 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:22 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:22 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:22 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.397527 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.405691 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" event={"ID":"cdf2c613-8426-4c4e-8cfe-bc015e17e47e","Type":"ContainerStarted","Data":"15b000b0f16a263df4954bb71725c8dd1489a665e7ae67a786f149b3d5a29911"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.430133 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xrlzc" event={"ID":"d64e4f2a-d704-4e63-992f-a6fe10c3ca87","Type":"ContainerStarted","Data":"3c40b0320dcca0a5fe8236151f88e0f0237100ed8dde7b515ec93d3afee6e682"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.459326 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:22 crc kubenswrapper[4943]: E1204 10:18:22.459974 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:22.959956155 +0000 UTC m=+151.548932023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.492028 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" event={"ID":"89aaae4b-99da-4820-b024-a013c629c63f","Type":"ContainerStarted","Data":"d4afd6e9fe14c419c72b4e0f735b1b62bf02a0f15f019f6fc8120b8fe07f1a6b"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.492384 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.524114 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" event={"ID":"7874756d-1da1-481a-be6e-df6034c4cac5","Type":"ContainerStarted","Data":"65e83d6cd7cbc5fe06d739d7baf9bfdca2ab0b88172048154add9a6024aa29d7"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.525317 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" podStartSLOduration=127.525298717 podStartE2EDuration="2m7.525298717s" podCreationTimestamp="2025-12-04 10:16:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:22.516985243 +0000 UTC m=+151.105961121" watchObservedRunningTime="2025-12-04 10:18:22.525298717 +0000 UTC m=+151.114274585" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.525585 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-xrlzc" podStartSLOduration=5.525580284 podStartE2EDuration="5.525580284s" podCreationTimestamp="2025-12-04 10:18:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:22.46423619 +0000 UTC m=+151.053212058" watchObservedRunningTime="2025-12-04 10:18:22.525580284 +0000 UTC m=+151.114556152" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.535527 4943 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-pxxxc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.535601 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" podUID="89aaae4b-99da-4820-b024-a013c629c63f" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.555960 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" event={"ID":"db1b9fc6-529d-4a5c-9ef9-b91b6f872225","Type":"ContainerStarted","Data":"07491cb89fe775593bb3b62d94e45fcb699a19d94efd7ccfe7b072fef7132040"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.560793 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:22 crc kubenswrapper[4943]: E1204 10:18:22.561083 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:23.061054083 +0000 UTC m=+151.650029961 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.609540 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" event={"ID":"149c7312-7be7-47d0-a2a4-bd0b0f588e3a","Type":"ContainerStarted","Data":"5060b3371fa9dc5d8cd287ec6b58572f7af09cfd3c002551d061fbc2ac2f9095"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.609596 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" event={"ID":"149c7312-7be7-47d0-a2a4-bd0b0f588e3a","Type":"ContainerStarted","Data":"1ad425c5edb3134c53ddf1222d8c77e7c9358d342f1956275d32fc8c23dc9cab"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.609757 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" event={"ID":"0c0b640d-f71b-4c0e-8b92-e592722ccee5","Type":"ContainerStarted","Data":"5c72210b0a53899af6448f53bdf2353071a15c8734be8132639c997404b7483b"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.644321 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-grw5p"] Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.658614 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" event={"ID":"3e8ecae4-2aee-4125-a16c-d4f340d840de","Type":"ContainerStarted","Data":"902274e1c69dbce2bd12daac8bad930ca8bcacdee955847d3fc0c0e36ed9ce80"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.665590 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.666145 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.666813 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" event={"ID":"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021","Type":"ContainerStarted","Data":"b44b5b13bd710ec52435fd3fe88e41da6efc83c8a09ee30d678e95148f96b908"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.667583 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ml4z6"] Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.671814 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-28rrg" event={"ID":"f93fafb8-e1d5-4ef2-a443-b3cfe19b2bda","Type":"ContainerStarted","Data":"52bd0e7fc39b96387dd9d41aebe9c8969cb58b492b339745b3092bac7f3efbee"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.680542 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:22 crc kubenswrapper[4943]: E1204 10:18:22.683627 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:23.183582325 +0000 UTC m=+151.772558193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.749437 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f"] Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.749880 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" event={"ID":"7ec6444d-9ad0-4310-974a-d8525e16e950","Type":"ContainerStarted","Data":"3af85b06a76d9c6fe15216851a6f17838dff68340db72b2af35aab3ac1d57bcd"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.789273 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" event={"ID":"d93e50fc-a2f7-4ed5-ad71-7590b393c71f","Type":"ContainerStarted","Data":"7f0937ba659415663423b15c0dd056ac9e409428eb4049312f2a9e4b0f84f650"} Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.789756 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:22 crc kubenswrapper[4943]: E1204 10:18:22.790155 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:23.290139467 +0000 UTC m=+151.879115335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.827362 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-6jx2n"] Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.890274 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-kdt7l"] Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.890669 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8nftj"] Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.891604 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:22 crc kubenswrapper[4943]: E1204 10:18:22.891963 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:23.391945932 +0000 UTC m=+151.980921800 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.906175 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2"] Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.906246 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-5bgm4"] Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.979987 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kvmzb" podStartSLOduration=128.979958628 podStartE2EDuration="2m8.979958628s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:22.968924148 +0000 UTC m=+151.557900036" watchObservedRunningTime="2025-12-04 10:18:22.979958628 +0000 UTC m=+151.568934496" Dec 04 10:18:22 crc kubenswrapper[4943]: I1204 10:18:22.992140 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:22 crc kubenswrapper[4943]: E1204 10:18:22.993185 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:23.493165572 +0000 UTC m=+152.082141450 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.057902 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-m4262"] Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.087619 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-rqdpv"] Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.094467 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:23 crc kubenswrapper[4943]: E1204 10:18:23.094906 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:23.594890335 +0000 UTC m=+152.183866203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.126992 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7"] Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.129894 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk"] Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.198334 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:23 crc kubenswrapper[4943]: E1204 10:18:23.198761 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:23.69873598 +0000 UTC m=+152.287711848 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.217888 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-fs4kj"] Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.249627 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5"] Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.253176 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:23 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:23 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:23 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.253275 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.299483 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:23 crc kubenswrapper[4943]: E1204 10:18:23.299800 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:23.799787286 +0000 UTC m=+152.388763154 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.389655 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd"] Dec 04 10:18:23 crc kubenswrapper[4943]: E1204 10:18:23.424349 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:23.924319766 +0000 UTC m=+152.513295634 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.432279 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.432757 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:23 crc kubenswrapper[4943]: E1204 10:18:23.433219 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:23.933184943 +0000 UTC m=+152.522160811 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.452777 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8"] Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.523218 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ph88v"] Dec 04 10:18:23 crc kubenswrapper[4943]: W1204 10:18:23.526368 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode188d23e_750e_4c02_9388_0c8bc4a2906d.slice/crio-3f94ad7c3e94fe90c2f699bb08d63caff7d9401cf508503d889c3fc80f50a209 WatchSource:0}: Error finding container 3f94ad7c3e94fe90c2f699bb08d63caff7d9401cf508503d889c3fc80f50a209: Status 404 returned error can't find the container with id 3f94ad7c3e94fe90c2f699bb08d63caff7d9401cf508503d889c3fc80f50a209 Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.539059 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:23 crc kubenswrapper[4943]: E1204 10:18:23.539444 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:24.039429818 +0000 UTC m=+152.628405686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.550927 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hgd88"] Dec 04 10:18:23 crc kubenswrapper[4943]: W1204 10:18:23.553110 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod560d2fc0_5e01_4d10_9499_72144de110f6.slice/crio-2082ba7c9399be35d292e40ed15fd88139436857442df11b2b778564c071eafc WatchSource:0}: Error finding container 2082ba7c9399be35d292e40ed15fd88139436857442df11b2b778564c071eafc: Status 404 returned error can't find the container with id 2082ba7c9399be35d292e40ed15fd88139436857442df11b2b778564c071eafc Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.640928 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:23 crc kubenswrapper[4943]: E1204 10:18:23.641349 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:24.141332395 +0000 UTC m=+152.730308263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.734048 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.742242 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:23 crc kubenswrapper[4943]: E1204 10:18:23.742587 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:24.242565035 +0000 UTC m=+152.831540903 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.836789 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kdt7l" event={"ID":"7848703f-6d8c-4915-8f42-e6222f2a889a","Type":"ContainerStarted","Data":"d6b821a61b57541b32605b1d23cd04c23378d818954f155bc6b318fcf0f06b04"} Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.843373 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:23 crc kubenswrapper[4943]: E1204 10:18:23.843737 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:24.343724164 +0000 UTC m=+152.932700032 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.912887 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-28rrg" event={"ID":"f93fafb8-e1d5-4ef2-a443-b3cfe19b2bda","Type":"ContainerStarted","Data":"9eb422fd31702f91c37c58096dc7d8149642d0a86eeedd1f9f676e476c23ee5c"} Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.915681 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" event={"ID":"3d93745e-ade6-45eb-b5e3-9b4270483319","Type":"ContainerStarted","Data":"c34b62251a66c74db535969420fd360b8158936227a98ac65367aa47e26de0c6"} Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.921329 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fs4kj" event={"ID":"d1a3d277-1dd7-4970-8017-4dca49e1b358","Type":"ContainerStarted","Data":"770fed10a2ae7f46e63c4d48f86ba375e342d8fdc85fc0f3477ab5bf396ccfce"} Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.933305 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" event={"ID":"db1b9fc6-529d-4a5c-9ef9-b91b6f872225","Type":"ContainerStarted","Data":"6ce972895f89ff9d013762ae67038d6b64f61399de83e88b5c34f6b04d085eb5"} Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.937742 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd" event={"ID":"5ef5c158-a074-47df-a4bb-2e1cab760745","Type":"ContainerStarted","Data":"0de5a4f0ac38eba28e67a80666c4ed52c3eb928f7a25b94ef0d27229301053fd"} Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.946715 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:23 crc kubenswrapper[4943]: E1204 10:18:23.947142 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:24.447125708 +0000 UTC m=+153.036101576 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:23 crc kubenswrapper[4943]: I1204 10:18:23.950962 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" event={"ID":"3e183218-08a1-4890-a6af-d56f3c7db2be","Type":"ContainerStarted","Data":"2fbb30276478cfd9c815dcae8d1122ebe4e49a5e5d750366871e0793a3eeb28a"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.006730 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-bnxzq" podStartSLOduration=129.006709068 podStartE2EDuration="2m9.006709068s" podCreationTimestamp="2025-12-04 10:16:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.004794382 +0000 UTC m=+152.593770250" watchObservedRunningTime="2025-12-04 10:18:24.006709068 +0000 UTC m=+152.595684936" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.024603 4943 generic.go:334] "Generic (PLEG): container finished" podID="d93e50fc-a2f7-4ed5-ad71-7590b393c71f" containerID="49254a1199f1c6b1609879f85853d7b5d0526cd12bd5be30c274b4b8af8382d9" exitCode=0 Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.024702 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" event={"ID":"d93e50fc-a2f7-4ed5-ad71-7590b393c71f","Type":"ContainerStarted","Data":"f876c15b32a6b5ff682e8f577a25b173759cac7b81befb6c734aab223e69e478"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.024729 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" event={"ID":"d93e50fc-a2f7-4ed5-ad71-7590b393c71f","Type":"ContainerDied","Data":"49254a1199f1c6b1609879f85853d7b5d0526cd12bd5be30c274b4b8af8382d9"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.025615 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.032798 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" event={"ID":"7874756d-1da1-481a-be6e-df6034c4cac5","Type":"ContainerStarted","Data":"889792af97881f93c461dab79073286d01b85ceb63cb07d5d257fc8d2482d68e"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.044615 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" event={"ID":"14bd93bd-8f38-48c3-9508-6c1ab9e65628","Type":"ContainerStarted","Data":"aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.045020 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.046619 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" event={"ID":"bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021","Type":"ContainerStarted","Data":"66d8944cc78101f7b872ce6271b82655b1af51e1350e142c3b18a8f9184c8e52"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.046619 4943 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-wkldv container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" start-of-body= Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.046866 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" podUID="14bd93bd-8f38-48c3-9508-6c1ab9e65628" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.13:6443/healthz\": dial tcp 10.217.0.13:6443: connect: connection refused" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.047416 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.048173 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.049120 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:24.549102608 +0000 UTC m=+153.138078486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.050244 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" event={"ID":"3e8ecae4-2aee-4125-a16c-d4f340d840de","Type":"ContainerStarted","Data":"92913f130aef3fe9cc66648ccd1f826469cbb0030c030049c570384e2f29e311"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.050725 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.050790 4943 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-dmb24 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.050813 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" podUID="bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.051594 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8488d27b42f9196ad526127baf44d57e96a0cd580d2cdb26b9259102194a5806"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.056439 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" event={"ID":"b3c8ac74-bfd7-494b-9181-36832b3ffa98","Type":"ContainerStarted","Data":"5490cbd199812c69ba28883c97905c6afd883af60fd3eb60f0d05808166467fb"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.065946 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-grw5p" event={"ID":"1abfb759-04f9-4885-b150-82355f8cf866","Type":"ContainerStarted","Data":"e4e7495b376f02a73fa8114035f14793967bd6380be70aa3e288017477bce72b"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.073943 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-grw5p" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.074102 4943 patch_prober.go:28] interesting pod/downloads-7954f5f757-grw5p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.074145 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grw5p" podUID="1abfb759-04f9-4885-b150-82355f8cf866" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.080600 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" event={"ID":"6e5550ab-55c3-4147-871f-3e49bf38ee55","Type":"ContainerStarted","Data":"788b79c4283662504c01f22099920c32d6f9584f03375bc572274f6c4acf69f8"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.081771 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" event={"ID":"e188d23e-750e-4c02-9388-0c8bc4a2906d","Type":"ContainerStarted","Data":"3f94ad7c3e94fe90c2f699bb08d63caff7d9401cf508503d889c3fc80f50a209"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.088601 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-m4262" event={"ID":"7f0e5e19-36d5-4553-896d-c5057c8fe1c3","Type":"ContainerStarted","Data":"d365de04b39db2e73745cf1a1385b14a81133e00f9dc3b92b10ada375219c298"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.092993 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" podStartSLOduration=131.092974412 podStartE2EDuration="2m11.092974412s" podCreationTimestamp="2025-12-04 10:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.06143333 +0000 UTC m=+152.650409198" watchObservedRunningTime="2025-12-04 10:18:24.092974412 +0000 UTC m=+152.681950290" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.097159 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b1d041164987315e8a62a9f4de36e0e276bf518ec1a6f442720ce2a55d489963"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.114318 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-779v8" podStartSLOduration=130.114299115 podStartE2EDuration="2m10.114299115s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.095534725 +0000 UTC m=+152.684510603" watchObservedRunningTime="2025-12-04 10:18:24.114299115 +0000 UTC m=+152.703274983" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.117156 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-rbhdx" podStartSLOduration=130.117144455 podStartE2EDuration="2m10.117144455s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.112600313 +0000 UTC m=+152.701576191" watchObservedRunningTime="2025-12-04 10:18:24.117144455 +0000 UTC m=+152.706120323" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.118589 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" event={"ID":"cdf2c613-8426-4c4e-8cfe-bc015e17e47e","Type":"ContainerStarted","Data":"312994f9b64c587b3f6b69ead8256b6dac2507913a51eb4163fa71986c997827"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.124617 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" event={"ID":"149c7312-7be7-47d0-a2a4-bd0b0f588e3a","Type":"ContainerStarted","Data":"94f100d2120bf58821cfeb963fe64f53db3f5da76e3acdd625e0e366f67ddeee"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.134487 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ml4z6" event={"ID":"7019146d-e522-4815-9642-9c9667aba409","Type":"ContainerStarted","Data":"3846299e076502bc6303eb1fa24e5a818c4e382fcd34ca2a44fd5d238fef03da"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.143285 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" event={"ID":"560d2fc0-5e01-4d10-9499-72144de110f6","Type":"ContainerStarted","Data":"2082ba7c9399be35d292e40ed15fd88139436857442df11b2b778564c071eafc"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.147865 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" podStartSLOduration=129.147847407 podStartE2EDuration="2m9.147847407s" podCreationTimestamp="2025-12-04 10:16:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.143522151 +0000 UTC m=+152.732498029" watchObservedRunningTime="2025-12-04 10:18:24.147847407 +0000 UTC m=+152.736823275" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.149151 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.150789 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"052ce50f5dc36b63ae8cd11fcc51312b3068a4103253f2bc5c4c41c379918315"} Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.151096 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:24.651074916 +0000 UTC m=+153.240050784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.151267 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.156723 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:24.656704634 +0000 UTC m=+153.245680502 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.158025 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7" event={"ID":"931858b0-187e-441e-a961-80c7cf44d916","Type":"ContainerStarted","Data":"e68ec18139855637a2424fd3260bcb3b2e2b88f2d417e04617542c75446761f7"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.159702 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" event={"ID":"cefa13a6-8806-4b35-9083-df5aa7d45f56","Type":"ContainerStarted","Data":"2e897b2c4054749bb9b25a1f85b836aa8a571273230f16b2dff2ee96d5e08538"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.161797 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" event={"ID":"b97e1699-6f5e-461d-b3e0-4afaef630174","Type":"ContainerStarted","Data":"a4d44af667fd48ed13f9688a723d63feb75210f29714172b6b697b9a2a011678"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.202612 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" podStartSLOduration=129.202589408 podStartE2EDuration="2m9.202589408s" podCreationTimestamp="2025-12-04 10:16:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.201324298 +0000 UTC m=+152.790300166" watchObservedRunningTime="2025-12-04 10:18:24.202589408 +0000 UTC m=+152.791565276" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.259887 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.264160 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:24.764133426 +0000 UTC m=+153.353109324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.264922 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" event={"ID":"7ec6444d-9ad0-4310-974a-d8525e16e950","Type":"ContainerStarted","Data":"7812eaa774b3fbfd47e550547630fa1c2e31fd5a8759ec62cf8d308bf5b9f609"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.268773 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:24 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:24 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:24 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.268852 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.323242 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hkrzs" podStartSLOduration=130.323199354 podStartE2EDuration="2m10.323199354s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.321525233 +0000 UTC m=+152.910501121" watchObservedRunningTime="2025-12-04 10:18:24.323199354 +0000 UTC m=+152.912175222" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.323475 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" podStartSLOduration=130.323470931 podStartE2EDuration="2m10.323470931s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.248319569 +0000 UTC m=+152.837295447" watchObservedRunningTime="2025-12-04 10:18:24.323470931 +0000 UTC m=+152.912446799" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.331618 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" event={"ID":"0e7f7c48-3d22-4aa4-8d06-f265758b3d36","Type":"ContainerStarted","Data":"30ea20b7a859d758b1f1b14fbfc835ba7e9887e16787a11c71566a2d9708bce4"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.358534 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" event={"ID":"c0352a66-8ffd-4949-ba0f-d8e6cb933ee4","Type":"ContainerStarted","Data":"8b4d6003f73f3b974ebeef3e52369d3efec6896a1c04c7392e50c751215e00f2"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.361403 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-x5dsb" podStartSLOduration=130.36138545 podStartE2EDuration="2m10.36138545s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.346613318 +0000 UTC m=+152.935589436" watchObservedRunningTime="2025-12-04 10:18:24.36138545 +0000 UTC m=+152.950361318" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.365269 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.365658 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:24.865644915 +0000 UTC m=+153.454620783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.377108 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" event={"ID":"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b","Type":"ContainerStarted","Data":"81222a94d2552265b9d6df71686e44e17414e3980db2f1752c9499839ec03fbb"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.398717 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-grw5p" podStartSLOduration=130.398694804 podStartE2EDuration="2m10.398694804s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.395946756 +0000 UTC m=+152.984922624" watchObservedRunningTime="2025-12-04 10:18:24.398694804 +0000 UTC m=+152.987670672" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.399011 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-tf5g6" event={"ID":"6c9f637d-2d43-4bb6-9db2-1831bd8e7573","Type":"ContainerStarted","Data":"7669a906c384643be868ffaf1b8fb049999b6eb9cd4162123fb21f4121a40f8c"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.399902 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.419124 4943 patch_prober.go:28] interesting pod/console-operator-58897d9998-tf5g6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.420029 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-tf5g6" podUID="6c9f637d-2d43-4bb6-9db2-1831bd8e7573" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.461356 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-6jx2n" event={"ID":"f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2","Type":"ContainerStarted","Data":"d60b619a9bf6f3b84ebec92de7bad7d85200673bfd2813ad6ee7767e3950c7b7"} Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.468983 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.471001 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:24.970975615 +0000 UTC m=+153.559951483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.484650 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h9rws" podStartSLOduration=131.484581679 podStartE2EDuration="2m11.484581679s" podCreationTimestamp="2025-12-04 10:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.445433089 +0000 UTC m=+153.034408957" watchObservedRunningTime="2025-12-04 10:18:24.484581679 +0000 UTC m=+153.073557557" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.504042 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-pxxxc" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.517508 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-4xwrm" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.532462 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" podStartSLOduration=130.532439321 podStartE2EDuration="2m10.532439321s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.485693106 +0000 UTC m=+153.074668974" watchObservedRunningTime="2025-12-04 10:18:24.532439321 +0000 UTC m=+153.121415189" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.539111 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-tf5g6" podStartSLOduration=131.539097045 podStartE2EDuration="2m11.539097045s" podCreationTimestamp="2025-12-04 10:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.531443537 +0000 UTC m=+153.120419415" watchObservedRunningTime="2025-12-04 10:18:24.539097045 +0000 UTC m=+153.128072923" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.570903 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.572170 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:25.072158174 +0000 UTC m=+153.661134042 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.585496 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-6jx2n" podStartSLOduration=7.58543582 podStartE2EDuration="7.58543582s" podCreationTimestamp="2025-12-04 10:18:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:24.552489802 +0000 UTC m=+153.141465670" watchObservedRunningTime="2025-12-04 10:18:24.58543582 +0000 UTC m=+153.174411688" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.672637 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.673241 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:25.17318597 +0000 UTC m=+153.762161838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.674167 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:25.174137804 +0000 UTC m=+153.763113682 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.673889 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.782034 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.782604 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:25.28257733 +0000 UTC m=+153.871553198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.782743 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.783194 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:25.283182746 +0000 UTC m=+153.872158614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.883864 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.884366 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:25.384346575 +0000 UTC m=+153.973322443 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:24 crc kubenswrapper[4943]: I1204 10:18:24.989406 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:24 crc kubenswrapper[4943]: E1204 10:18:24.990229 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:25.490181148 +0000 UTC m=+154.079157016 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.091938 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:25 crc kubenswrapper[4943]: E1204 10:18:25.092784 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:25.592760831 +0000 UTC m=+154.181736699 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.195524 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:25 crc kubenswrapper[4943]: E1204 10:18:25.195944 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:25.69592773 +0000 UTC m=+154.284903598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.249614 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:25 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:25 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:25 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.250397 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.309754 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:25 crc kubenswrapper[4943]: E1204 10:18:25.310564 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:25.810545189 +0000 UTC m=+154.399521057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.413827 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:25 crc kubenswrapper[4943]: E1204 10:18:25.414551 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:25.914536726 +0000 UTC m=+154.503512594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.505024 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"36b23f69cad43c11626aeaf701706ab66f8c855a6b50c54b28dfc9c5d2fefeaa"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.506421 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.522444 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:25 crc kubenswrapper[4943]: E1204 10:18:25.522759 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:26.022742549 +0000 UTC m=+154.611718417 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.541441 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2ec9c118a7fa8a1812b82124625a4dcafd264d6522646bacec7bbdd378ad8e9a"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.548941 4943 generic.go:334] "Generic (PLEG): container finished" podID="a3b7dc03-7238-49ce-8ad6-1eabd0198a87" containerID="02e6fee0a87f0fa26073935f809b8b393c46a4c1abd28554f89117833a8d6e9d" exitCode=0 Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.549047 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" event={"ID":"a3b7dc03-7238-49ce-8ad6-1eabd0198a87","Type":"ContainerDied","Data":"02e6fee0a87f0fa26073935f809b8b393c46a4c1abd28554f89117833a8d6e9d"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.606421 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-kdt7l" event={"ID":"7848703f-6d8c-4915-8f42-e6222f2a889a","Type":"ContainerStarted","Data":"1a4d16969d2172d65daa27c0b7c8c6424299b3c4287dffce1f85c17ffc1532ea"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.623645 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:25 crc kubenswrapper[4943]: E1204 10:18:25.624053 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:26.124037241 +0000 UTC m=+154.713013109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.646511 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" event={"ID":"e188d23e-750e-4c02-9388-0c8bc4a2906d","Type":"ContainerStarted","Data":"1c21d6092323145bf9a51f50e6b93f1e58d2f10d6dfe015945279e630e9d727a"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.646549 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.654768 4943 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-fzpv5 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.654826 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" podUID="e188d23e-750e-4c02-9388-0c8bc4a2906d" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.673908 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-kdt7l" podStartSLOduration=131.673862592 podStartE2EDuration="2m11.673862592s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:25.673765829 +0000 UTC m=+154.262741697" watchObservedRunningTime="2025-12-04 10:18:25.673862592 +0000 UTC m=+154.262838460" Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.674915 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" event={"ID":"6e5550ab-55c3-4147-871f-3e49bf38ee55","Type":"ContainerStarted","Data":"f5bb35adfa537bc9451bb24f0eee4eea074fa5103cdd7e442ebf6324e19ee278"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.765111 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:25 crc kubenswrapper[4943]: E1204 10:18:25.766562 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:26.266541303 +0000 UTC m=+154.855517181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.770618 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" podStartSLOduration=130.770600812 podStartE2EDuration="2m10.770600812s" podCreationTimestamp="2025-12-04 10:16:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:25.768772717 +0000 UTC m=+154.357748585" watchObservedRunningTime="2025-12-04 10:18:25.770600812 +0000 UTC m=+154.359576680" Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.774501 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2f3d0e53931c18c4813fef3872458ef6420cc7635821d7d11e24aa6c81402915"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.792471 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-6jx2n" event={"ID":"f3f7deb7-bd4f-40c7-b7d3-a7f78a7d00f2","Type":"ContainerStarted","Data":"eaab426bf52f95adfe913ba1e3030ed6264a79e79cb7066524238c8ccc2f9c9b"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.807471 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-hjcd2" podStartSLOduration=131.807447095 podStartE2EDuration="2m11.807447095s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:25.798049035 +0000 UTC m=+154.387024913" watchObservedRunningTime="2025-12-04 10:18:25.807447095 +0000 UTC m=+154.396422963" Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.810524 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" event={"ID":"cefa13a6-8806-4b35-9083-df5aa7d45f56","Type":"ContainerStarted","Data":"12f614b8f841c4b9a4714f078170871b2778971fdbaf42aef177a87214c4f61d"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.849919 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-28rrg" event={"ID":"f93fafb8-e1d5-4ef2-a443-b3cfe19b2bda","Type":"ContainerStarted","Data":"633417d15869b194f7e2d7d6d44f3b28ea52f2c2259146b566af7faf5760b8ef"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.866671 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:25 crc kubenswrapper[4943]: E1204 10:18:25.867401 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:26.367383704 +0000 UTC m=+154.956359572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.867853 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" event={"ID":"3d93745e-ade6-45eb-b5e3-9b4270483319","Type":"ContainerStarted","Data":"ef80f77f610c6c38afaa8be5fa8bfbee3f93e1ddbf206ddfaa75e16c46b4c27d"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.893118 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ml4z6" event={"ID":"7019146d-e522-4815-9642-9c9667aba409","Type":"ContainerStarted","Data":"4e3b43a398979115ffe1d6eff622db7a25f9ef6dd7f74b93146ea5d86f27610b"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.894901 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-snz6f" podStartSLOduration=131.894875477 podStartE2EDuration="2m11.894875477s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:25.8933424 +0000 UTC m=+154.482318268" watchObservedRunningTime="2025-12-04 10:18:25.894875477 +0000 UTC m=+154.483851335" Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.897404 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd" event={"ID":"5ef5c158-a074-47df-a4bb-2e1cab760745","Type":"ContainerStarted","Data":"3d55856c286b19a16641d7af76357a3af1ad30160db501baee8cbc9b19766f25"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.973772 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:25 crc kubenswrapper[4943]: E1204 10:18:25.974727 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:26.474707914 +0000 UTC m=+155.063683782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.980455 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-grw5p" event={"ID":"1abfb759-04f9-4885-b150-82355f8cf866","Type":"ContainerStarted","Data":"fa526d53352cd299c51d6d48741e1405a730a760a89a472c9f9fc9cb2e561764"} Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.992495 4943 patch_prober.go:28] interesting pod/downloads-7954f5f757-grw5p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 04 10:18:25 crc kubenswrapper[4943]: I1204 10:18:25.992862 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grw5p" podUID="1abfb759-04f9-4885-b150-82355f8cf866" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.001336 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-28rrg" podStartSLOduration=132.001302175 podStartE2EDuration="2m12.001302175s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:25.998711182 +0000 UTC m=+154.587687050" watchObservedRunningTime="2025-12-04 10:18:26.001302175 +0000 UTC m=+154.590278043" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.032856 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" event={"ID":"3e183218-08a1-4890-a6af-d56f3c7db2be","Type":"ContainerStarted","Data":"698a6d0e6901b22f5d4757974553cd340604d29a3013380b7eb9ce2eae17e20e"} Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.033352 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.034601 4943 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ph88v container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.034647 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" podUID="3e183218-08a1-4890-a6af-d56f3c7db2be" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.035662 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" event={"ID":"b97e1699-6f5e-461d-b3e0-4afaef630174","Type":"ContainerStarted","Data":"b16e8939a07c1b6836562a4960375b3adb26f6b3fa099dfa2e8a86444cf9497d"} Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.037571 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8nftj" event={"ID":"0e7f7c48-3d22-4aa4-8d06-f265758b3d36","Type":"ContainerStarted","Data":"f059bb15757ef4b41734f5352057606be72ad4e35496139b7a7dd5d02c3a9418"} Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.039727 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" event={"ID":"c0352a66-8ffd-4949-ba0f-d8e6cb933ee4","Type":"ContainerStarted","Data":"3ee7f2e66030b9ff6b906e5f4c93f01bc2280f685a949940462b12c991ab69a1"} Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.042103 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7" event={"ID":"931858b0-187e-441e-a961-80c7cf44d916","Type":"ContainerStarted","Data":"4906ce765f2ea5830845bcc3b722e34c7327d6a3c08a5b3c68b39efdcef28237"} Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.100007 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:26 crc kubenswrapper[4943]: E1204 10:18:26.101086 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:26.60107221 +0000 UTC m=+155.190048078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.129119 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" event={"ID":"560d2fc0-5e01-4d10-9499-72144de110f6","Type":"ContainerStarted","Data":"c06ba320d501cf06e57f3ce19cfb06a9f263b8eb9f91f08d31f929953c620ffe"} Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.152082 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-tf5g6" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.161815 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.163515 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7" podStartSLOduration=133.16349625 podStartE2EDuration="2m13.16349625s" podCreationTimestamp="2025-12-04 10:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:26.161783568 +0000 UTC m=+154.750759446" watchObservedRunningTime="2025-12-04 10:18:26.16349625 +0000 UTC m=+154.752472118" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.164788 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-vhxkd" podStartSLOduration=132.164779672 podStartE2EDuration="2m12.164779672s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:26.10190075 +0000 UTC m=+154.690876618" watchObservedRunningTime="2025-12-04 10:18:26.164779672 +0000 UTC m=+154.753755540" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.248846 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:26 crc kubenswrapper[4943]: E1204 10:18:26.250776 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:26.750748809 +0000 UTC m=+155.339724677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.251460 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:26 crc kubenswrapper[4943]: E1204 10:18:26.251722 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:26.751710632 +0000 UTC m=+155.340686500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.253497 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:26 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:26 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:26 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.253528 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.268141 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-89wdj" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.310281 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-cm2tk" podStartSLOduration=132.310260047 podStartE2EDuration="2m12.310260047s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:26.305685544 +0000 UTC m=+154.894661412" watchObservedRunningTime="2025-12-04 10:18:26.310260047 +0000 UTC m=+154.899235915" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.376351 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:26 crc kubenswrapper[4943]: E1204 10:18:26.377356 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:26.87733638 +0000 UTC m=+155.466312238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.401751 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hgd88" podStartSLOduration=131.401723757 podStartE2EDuration="2m11.401723757s" podCreationTimestamp="2025-12-04 10:16:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:26.343888911 +0000 UTC m=+154.932864799" watchObservedRunningTime="2025-12-04 10:18:26.401723757 +0000 UTC m=+154.990699625" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.403440 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" podStartSLOduration=132.403424379 podStartE2EDuration="2m12.403424379s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:26.374025369 +0000 UTC m=+154.963001257" watchObservedRunningTime="2025-12-04 10:18:26.403424379 +0000 UTC m=+154.992400257" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.478571 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:26 crc kubenswrapper[4943]: E1204 10:18:26.479300 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:26.979283398 +0000 UTC m=+155.568259266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.580884 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:26 crc kubenswrapper[4943]: E1204 10:18:26.581228 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:27.081183385 +0000 UTC m=+155.670159253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.581409 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.581772 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" podStartSLOduration=132.581737009 podStartE2EDuration="2m12.581737009s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:26.580897158 +0000 UTC m=+155.169873026" watchObservedRunningTime="2025-12-04 10:18:26.581737009 +0000 UTC m=+155.170712877" Dec 04 10:18:26 crc kubenswrapper[4943]: E1204 10:18:26.581844 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:27.081835982 +0000 UTC m=+155.670811850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.686626 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:26 crc kubenswrapper[4943]: E1204 10:18:26.686979 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:27.186958257 +0000 UTC m=+155.775934125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:26 crc kubenswrapper[4943]: I1204 10:18:26.925536 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:26 crc kubenswrapper[4943]: E1204 10:18:26.925950 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:27.425933264 +0000 UTC m=+156.014909132 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.025961 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:27 crc kubenswrapper[4943]: E1204 10:18:27.029213 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:27.529178303 +0000 UTC m=+156.118154161 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.128639 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:27 crc kubenswrapper[4943]: E1204 10:18:27.129083 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:27.629066901 +0000 UTC m=+156.218042769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.138443 4943 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-dmb24 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.138503 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" podUID="bd5bfbb1-0dfd-4925-bfe3-827dc3c6e021" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.180404 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fs4kj" event={"ID":"d1a3d277-1dd7-4970-8017-4dca49e1b358","Type":"ContainerStarted","Data":"b48510577e1a96a4c8ba843ae9b7e7cf9574328e0a79eb57a6dc5fdea0f70880"} Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.180458 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fs4kj" event={"ID":"d1a3d277-1dd7-4970-8017-4dca49e1b358","Type":"ContainerStarted","Data":"a8bbdefa5212870edcf7c2a69671a94b613c8cab76e00eccf487c3867dd57ff5"} Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.181391 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.193887 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ml4z6" event={"ID":"7019146d-e522-4815-9642-9c9667aba409","Type":"ContainerStarted","Data":"aca733aaf75b79be03160548aee0af45ea2ee949c7a660bb2d9ae10f9d4caf46"} Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.198069 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fqmc8" event={"ID":"560d2fc0-5e01-4d10-9499-72144de110f6","Type":"ContainerStarted","Data":"7cfa34fc37306b11f4eabfec2c49f13ee2984a2f53e2a96ed78e8ef7268887aa"} Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.218159 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-fs4kj" podStartSLOduration=10.218137264 podStartE2EDuration="10.218137264s" podCreationTimestamp="2025-12-04 10:18:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:27.214502995 +0000 UTC m=+155.803478873" watchObservedRunningTime="2025-12-04 10:18:27.218137264 +0000 UTC m=+155.807113142" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.220628 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" event={"ID":"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b","Type":"ContainerStarted","Data":"3b6f11a0450b7dee3c98213d9859479a7cc7ef9a3bda8e2a1e956a993cbe65c2"} Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.226706 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-cx7l7" event={"ID":"931858b0-187e-441e-a961-80c7cf44d916","Type":"ContainerStarted","Data":"3fb9701aff089e26667184a66b0d2e2a96c59b263ff885c4b5a9161f2e4f76ac"} Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.230097 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:27 crc kubenswrapper[4943]: E1204 10:18:27.230668 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:27.73064002 +0000 UTC m=+156.319615888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.247992 4943 generic.go:334] "Generic (PLEG): container finished" podID="3d93745e-ade6-45eb-b5e3-9b4270483319" containerID="ef80f77f610c6c38afaa8be5fa8bfbee3f93e1ddbf206ddfaa75e16c46b4c27d" exitCode=0 Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.248088 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" event={"ID":"3d93745e-ade6-45eb-b5e3-9b4270483319","Type":"ContainerDied","Data":"ef80f77f610c6c38afaa8be5fa8bfbee3f93e1ddbf206ddfaa75e16c46b4c27d"} Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.248122 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" event={"ID":"3d93745e-ade6-45eb-b5e3-9b4270483319","Type":"ContainerStarted","Data":"a7f196c677c172c728a9825ad755e6bb07c789c80d57c69f7539b33bccedd2ec"} Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.249310 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-ml4z6" podStartSLOduration=133.249296978 podStartE2EDuration="2m13.249296978s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:27.248318184 +0000 UTC m=+155.837294052" watchObservedRunningTime="2025-12-04 10:18:27.249296978 +0000 UTC m=+155.838272846" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.255554 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:27 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:27 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:27 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.255614 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.269606 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-m4262" event={"ID":"7f0e5e19-36d5-4553-896d-c5057c8fe1c3","Type":"ContainerStarted","Data":"1cc6cabb22a6fbb53186b9552ea802b8a0ada2d7c9a507c4f017f3f9bcb984cf"} Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.269655 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-m4262" event={"ID":"7f0e5e19-36d5-4553-896d-c5057c8fe1c3","Type":"ContainerStarted","Data":"bde11a58512cfd41359f7b9e168bc28b04183e166034f44a9fd085fe1b6a9cb5"} Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.271226 4943 patch_prober.go:28] interesting pod/downloads-7954f5f757-grw5p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.271310 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grw5p" podUID="1abfb759-04f9-4885-b150-82355f8cf866" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.271611 4943 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ph88v container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.271631 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" podUID="3e183218-08a1-4890-a6af-d56f3c7db2be" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.288975 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-fzpv5" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.290446 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-dmb24" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.304532 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-m4262" podStartSLOduration=133.304511521 podStartE2EDuration="2m13.304511521s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:27.303719831 +0000 UTC m=+155.892695719" watchObservedRunningTime="2025-12-04 10:18:27.304511521 +0000 UTC m=+155.893487399" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.357003 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:27 crc kubenswrapper[4943]: E1204 10:18:27.383585 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:27.883542828 +0000 UTC m=+156.472518696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.484554 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:27 crc kubenswrapper[4943]: E1204 10:18:27.485165 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:27.985105396 +0000 UTC m=+156.574081274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.587191 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:27 crc kubenswrapper[4943]: E1204 10:18:27.587638 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:28.087621949 +0000 UTC m=+156.676597817 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.688666 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:27 crc kubenswrapper[4943]: E1204 10:18:27.689142 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:28.189120756 +0000 UTC m=+156.778096624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.790571 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:27 crc kubenswrapper[4943]: E1204 10:18:27.791018 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:28.291000292 +0000 UTC m=+156.879976200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.821171 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.893410 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-config-volume\") pod \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.893544 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nfln\" (UniqueName: \"kubernetes.io/projected/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-kube-api-access-7nfln\") pod \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.894332 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-config-volume" (OuterVolumeSpecName: "config-volume") pod "a3b7dc03-7238-49ce-8ad6-1eabd0198a87" (UID: "a3b7dc03-7238-49ce-8ad6-1eabd0198a87"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.895035 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.895095 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-secret-volume\") pod \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\" (UID: \"a3b7dc03-7238-49ce-8ad6-1eabd0198a87\") " Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.895546 4943 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 10:18:27 crc kubenswrapper[4943]: E1204 10:18:27.895983 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:28.395957614 +0000 UTC m=+156.984933482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.901767 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-kube-api-access-7nfln" (OuterVolumeSpecName: "kube-api-access-7nfln") pod "a3b7dc03-7238-49ce-8ad6-1eabd0198a87" (UID: "a3b7dc03-7238-49ce-8ad6-1eabd0198a87"). InnerVolumeSpecName "kube-api-access-7nfln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.908397 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a3b7dc03-7238-49ce-8ad6-1eabd0198a87" (UID: "a3b7dc03-7238-49ce-8ad6-1eabd0198a87"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.997125 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.997311 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nfln\" (UniqueName: \"kubernetes.io/projected/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-kube-api-access-7nfln\") on node \"crc\" DevicePath \"\"" Dec 04 10:18:27 crc kubenswrapper[4943]: I1204 10:18:27.997327 4943 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3b7dc03-7238-49ce-8ad6-1eabd0198a87-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 10:18:27 crc kubenswrapper[4943]: E1204 10:18:27.997674 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:28.497660156 +0000 UTC m=+157.086636024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.028609 4943 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.098497 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:28 crc kubenswrapper[4943]: E1204 10:18:28.098814 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:28.598798845 +0000 UTC m=+157.187774713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.200309 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:28 crc kubenswrapper[4943]: E1204 10:18:28.201015 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:28.7010021 +0000 UTC m=+157.289977958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.244349 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:28 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:28 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:28 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.244422 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.304970 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:28 crc kubenswrapper[4943]: E1204 10:18:28.305161 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 10:18:28.805130691 +0000 UTC m=+157.394106559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.305286 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:28 crc kubenswrapper[4943]: E1204 10:18:28.305597 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 10:18:28.805589463 +0000 UTC m=+157.394565331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wmxwg" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.318363 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" event={"ID":"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b","Type":"ContainerStarted","Data":"a05f185c9a0891ee6c1c5932bd74eee03a6907cfc7d050aa19488724274297aa"} Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.318428 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" event={"ID":"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b","Type":"ContainerStarted","Data":"4867c45041be06c871c6d40483240528f243476481fb02261eacbf9fb50a7e22"} Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.326948 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" event={"ID":"a3b7dc03-7238-49ce-8ad6-1eabd0198a87","Type":"ContainerDied","Data":"1760c0af72bde1fa103a90b058e355793f1c07d1f45e33f9efe4687bca957c2d"} Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.327020 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1760c0af72bde1fa103a90b058e355793f1c07d1f45e33f9efe4687bca957c2d" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.327125 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414055-z8m4l" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.342341 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" event={"ID":"3d93745e-ade6-45eb-b5e3-9b4270483319","Type":"ContainerStarted","Data":"6085d147c5a0ebbe15e427782170bee1e7defa50a07453378db86ee5e432678e"} Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.343824 4943 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ph88v container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.343871 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" podUID="3e183218-08a1-4890-a6af-d56f3c7db2be" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.356452 4943 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-04T10:18:28.028642976Z","Handler":null,"Name":""} Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.404241 4943 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.404305 4943 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.406315 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.427817 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.508375 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.550063 4943 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.550123 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.571041 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.665673 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" podStartSLOduration=134.665641726 podStartE2EDuration="2m14.665641726s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:28.440957149 +0000 UTC m=+157.029933007" watchObservedRunningTime="2025-12-04 10:18:28.665641726 +0000 UTC m=+157.254617594" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.667290 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jcsk7"] Dec 04 10:18:28 crc kubenswrapper[4943]: E1204 10:18:28.667632 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3b7dc03-7238-49ce-8ad6-1eabd0198a87" containerName="collect-profiles" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.667664 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3b7dc03-7238-49ce-8ad6-1eabd0198a87" containerName="collect-profiles" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.667771 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3b7dc03-7238-49ce-8ad6-1eabd0198a87" containerName="collect-profiles" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.668732 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.672141 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.710839 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d22ad643-0a77-49fb-a7f1-54603fe0f82d-catalog-content\") pod \"community-operators-jcsk7\" (UID: \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\") " pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.711175 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfjpc\" (UniqueName: \"kubernetes.io/projected/d22ad643-0a77-49fb-a7f1-54603fe0f82d-kube-api-access-rfjpc\") pod \"community-operators-jcsk7\" (UID: \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\") " pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.711297 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d22ad643-0a77-49fb-a7f1-54603fe0f82d-utilities\") pod \"community-operators-jcsk7\" (UID: \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\") " pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.723041 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wmxwg\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.729259 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jcsk7"] Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.791264 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.812831 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d22ad643-0a77-49fb-a7f1-54603fe0f82d-catalog-content\") pod \"community-operators-jcsk7\" (UID: \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\") " pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.812919 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfjpc\" (UniqueName: \"kubernetes.io/projected/d22ad643-0a77-49fb-a7f1-54603fe0f82d-kube-api-access-rfjpc\") pod \"community-operators-jcsk7\" (UID: \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\") " pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.813072 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d22ad643-0a77-49fb-a7f1-54603fe0f82d-utilities\") pod \"community-operators-jcsk7\" (UID: \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\") " pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.814629 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d22ad643-0a77-49fb-a7f1-54603fe0f82d-catalog-content\") pod \"community-operators-jcsk7\" (UID: \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\") " pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.814694 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d22ad643-0a77-49fb-a7f1-54603fe0f82d-utilities\") pod \"community-operators-jcsk7\" (UID: \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\") " pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.846238 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfjpc\" (UniqueName: \"kubernetes.io/projected/d22ad643-0a77-49fb-a7f1-54603fe0f82d-kube-api-access-rfjpc\") pod \"community-operators-jcsk7\" (UID: \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\") " pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.868268 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l7p8s"] Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.870422 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.873675 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l7p8s"] Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.876366 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.916822 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-utilities\") pod \"certified-operators-l7p8s\" (UID: \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\") " pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.916963 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwhw9\" (UniqueName: \"kubernetes.io/projected/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-kube-api-access-gwhw9\") pod \"certified-operators-l7p8s\" (UID: \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\") " pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.917002 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-catalog-content\") pod \"certified-operators-l7p8s\" (UID: \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\") " pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:18:28 crc kubenswrapper[4943]: I1204 10:18:28.990600 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.017766 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwhw9\" (UniqueName: \"kubernetes.io/projected/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-kube-api-access-gwhw9\") pod \"certified-operators-l7p8s\" (UID: \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\") " pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.017823 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-catalog-content\") pod \"certified-operators-l7p8s\" (UID: \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\") " pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.017849 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-utilities\") pod \"certified-operators-l7p8s\" (UID: \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\") " pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.018351 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-utilities\") pod \"certified-operators-l7p8s\" (UID: \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\") " pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.018625 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-catalog-content\") pod \"certified-operators-l7p8s\" (UID: \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\") " pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.043180 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwhw9\" (UniqueName: \"kubernetes.io/projected/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-kube-api-access-gwhw9\") pod \"certified-operators-l7p8s\" (UID: \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\") " pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.118725 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lkx78"] Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.119897 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.195562 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lkx78"] Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.220152 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88vlp\" (UniqueName: \"kubernetes.io/projected/7462ea89-12b3-4d4a-9757-b2e97b190a53-kube-api-access-88vlp\") pod \"community-operators-lkx78\" (UID: \"7462ea89-12b3-4d4a-9757-b2e97b190a53\") " pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.220275 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7462ea89-12b3-4d4a-9757-b2e97b190a53-catalog-content\") pod \"community-operators-lkx78\" (UID: \"7462ea89-12b3-4d4a-9757-b2e97b190a53\") " pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.220334 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7462ea89-12b3-4d4a-9757-b2e97b190a53-utilities\") pod \"community-operators-lkx78\" (UID: \"7462ea89-12b3-4d4a-9757-b2e97b190a53\") " pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.234073 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.260212 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:29 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:29 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:29 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.260297 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.354729 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88vlp\" (UniqueName: \"kubernetes.io/projected/7462ea89-12b3-4d4a-9757-b2e97b190a53-kube-api-access-88vlp\") pod \"community-operators-lkx78\" (UID: \"7462ea89-12b3-4d4a-9757-b2e97b190a53\") " pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.355190 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7462ea89-12b3-4d4a-9757-b2e97b190a53-catalog-content\") pod \"community-operators-lkx78\" (UID: \"7462ea89-12b3-4d4a-9757-b2e97b190a53\") " pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.355307 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7462ea89-12b3-4d4a-9757-b2e97b190a53-utilities\") pod \"community-operators-lkx78\" (UID: \"7462ea89-12b3-4d4a-9757-b2e97b190a53\") " pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.356255 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7462ea89-12b3-4d4a-9757-b2e97b190a53-utilities\") pod \"community-operators-lkx78\" (UID: \"7462ea89-12b3-4d4a-9757-b2e97b190a53\") " pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.356810 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b8cs7"] Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.357946 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7462ea89-12b3-4d4a-9757-b2e97b190a53-catalog-content\") pod \"community-operators-lkx78\" (UID: \"7462ea89-12b3-4d4a-9757-b2e97b190a53\") " pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.372374 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b8cs7"] Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.372511 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.408701 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88vlp\" (UniqueName: \"kubernetes.io/projected/7462ea89-12b3-4d4a-9757-b2e97b190a53-kube-api-access-88vlp\") pod \"community-operators-lkx78\" (UID: \"7462ea89-12b3-4d4a-9757-b2e97b190a53\") " pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.421520 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" event={"ID":"d9c45054-0f0f-4bc4-9efd-5b45c98dda8b","Type":"ContainerStarted","Data":"10ede81dc27ace3a5eda7a6446f8c4ba17f6770c6a7d20c50f88c2361673292a"} Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.459482 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-utilities\") pod \"certified-operators-b8cs7\" (UID: \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\") " pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.459620 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wzkd\" (UniqueName: \"kubernetes.io/projected/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-kube-api-access-8wzkd\") pod \"certified-operators-b8cs7\" (UID: \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\") " pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.459675 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-catalog-content\") pod \"certified-operators-b8cs7\" (UID: \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\") " pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.470701 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.487835 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-rqdpv" podStartSLOduration=12.487812043 podStartE2EDuration="12.487812043s" podCreationTimestamp="2025-12-04 10:18:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:29.486754307 +0000 UTC m=+158.075730175" watchObservedRunningTime="2025-12-04 10:18:29.487812043 +0000 UTC m=+158.076787921" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.588449 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wzkd\" (UniqueName: \"kubernetes.io/projected/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-kube-api-access-8wzkd\") pod \"certified-operators-b8cs7\" (UID: \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\") " pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.588529 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-catalog-content\") pod \"certified-operators-b8cs7\" (UID: \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\") " pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.588612 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-utilities\") pod \"certified-operators-b8cs7\" (UID: \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\") " pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.589626 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-catalog-content\") pod \"certified-operators-b8cs7\" (UID: \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\") " pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.591123 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-utilities\") pod \"certified-operators-b8cs7\" (UID: \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\") " pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.638876 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wzkd\" (UniqueName: \"kubernetes.io/projected/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-kube-api-access-8wzkd\") pod \"certified-operators-b8cs7\" (UID: \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\") " pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.672925 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.673225 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.702467 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wmxwg"] Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.721635 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:18:29 crc kubenswrapper[4943]: W1204 10:18:29.724635 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6da7528_9b01_40b4_88d3_e78a3c178300.slice/crio-c3e3832740fa9bee9196c6115d250312879a88b42171b0e5d88dc4f82fb57c8a WatchSource:0}: Error finding container c3e3832740fa9bee9196c6115d250312879a88b42171b0e5d88dc4f82fb57c8a: Status 404 returned error can't find the container with id c3e3832740fa9bee9196c6115d250312879a88b42171b0e5d88dc4f82fb57c8a Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.871494 4943 patch_prober.go:28] interesting pod/downloads-7954f5f757-grw5p container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.871535 4943 patch_prober.go:28] interesting pod/downloads-7954f5f757-grw5p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.871577 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-grw5p" podUID="1abfb759-04f9-4885-b150-82355f8cf866" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.871618 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grw5p" podUID="1abfb759-04f9-4885-b150-82355f8cf866" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.988882 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.989839 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.995831 4943 patch_prober.go:28] interesting pod/console-f9d7485db-kdt7l container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 04 10:18:29 crc kubenswrapper[4943]: I1204 10:18:29.995892 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-kdt7l" podUID="7848703f-6d8c-4915-8f42-e6222f2a889a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.055560 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jcsk7"] Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.060893 4943 patch_prober.go:28] interesting pod/apiserver-76f77b778f-5bgm4 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 04 10:18:30 crc kubenswrapper[4943]: [+]log ok Dec 04 10:18:30 crc kubenswrapper[4943]: [+]etcd ok Dec 04 10:18:30 crc kubenswrapper[4943]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 04 10:18:30 crc kubenswrapper[4943]: [+]poststarthook/generic-apiserver-start-informers ok Dec 04 10:18:30 crc kubenswrapper[4943]: [+]poststarthook/max-in-flight-filter ok Dec 04 10:18:30 crc kubenswrapper[4943]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 04 10:18:30 crc kubenswrapper[4943]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 04 10:18:30 crc kubenswrapper[4943]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 04 10:18:30 crc kubenswrapper[4943]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 04 10:18:30 crc kubenswrapper[4943]: [+]poststarthook/project.openshift.io-projectcache ok Dec 04 10:18:30 crc kubenswrapper[4943]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 04 10:18:30 crc kubenswrapper[4943]: [+]poststarthook/openshift.io-startinformers ok Dec 04 10:18:30 crc kubenswrapper[4943]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 04 10:18:30 crc kubenswrapper[4943]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 04 10:18:30 crc kubenswrapper[4943]: livez check failed Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.060999 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" podUID="3d93745e-ade6-45eb-b5e3-9b4270483319" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.096031 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l7p8s"] Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.108527 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lkx78"] Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.337780 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.340930 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b8cs7"] Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.341924 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:30 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:30 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:30 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.341985 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.430256 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8cs7" event={"ID":"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565","Type":"ContainerStarted","Data":"837a4a83d35517c20078f0044d61279b17c3c40925e9bd000baaa9b7fc49e565"} Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.431827 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkx78" event={"ID":"7462ea89-12b3-4d4a-9757-b2e97b190a53","Type":"ContainerStarted","Data":"e70325153e732945492149759474892083912a69056da658c340e603e31a3937"} Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.432709 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7p8s" event={"ID":"0d555ec7-2800-45a7-a1a2-27b1a435b0d4","Type":"ContainerStarted","Data":"3cdac60ea1b14a02105a384481bc8c94b501b757ee13247fa16dae1703d0de2a"} Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.433577 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" event={"ID":"a6da7528-9b01-40b4-88d3-e78a3c178300","Type":"ContainerStarted","Data":"c3e3832740fa9bee9196c6115d250312879a88b42171b0e5d88dc4f82fb57c8a"} Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.436049 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jcsk7" event={"ID":"d22ad643-0a77-49fb-a7f1-54603fe0f82d","Type":"ContainerStarted","Data":"1c52d974a54ca301b60ff950116233e35c8fcfb42112157e236698a1cc91b7dd"} Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.447587 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.950300 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q9rd8"] Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.952327 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.956127 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 10:18:30 crc kubenswrapper[4943]: I1204 10:18:30.957476 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9rd8"] Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.101988 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99lx2\" (UniqueName: \"kubernetes.io/projected/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-kube-api-access-99lx2\") pod \"redhat-marketplace-q9rd8\" (UID: \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\") " pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.102516 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-utilities\") pod \"redhat-marketplace-q9rd8\" (UID: \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\") " pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.103464 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-catalog-content\") pod \"redhat-marketplace-q9rd8\" (UID: \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\") " pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.204663 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99lx2\" (UniqueName: \"kubernetes.io/projected/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-kube-api-access-99lx2\") pod \"redhat-marketplace-q9rd8\" (UID: \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\") " pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.204703 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-utilities\") pod \"redhat-marketplace-q9rd8\" (UID: \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\") " pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.205022 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-catalog-content\") pod \"redhat-marketplace-q9rd8\" (UID: \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\") " pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.205105 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-utilities\") pod \"redhat-marketplace-q9rd8\" (UID: \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\") " pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.205491 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-catalog-content\") pod \"redhat-marketplace-q9rd8\" (UID: \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\") " pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.235226 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99lx2\" (UniqueName: \"kubernetes.io/projected/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-kube-api-access-99lx2\") pod \"redhat-marketplace-q9rd8\" (UID: \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\") " pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.240152 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5cpfp"] Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.241470 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.252181 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:31 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:31 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:31 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.252558 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.252602 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5cpfp"] Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.358498 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.359355 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.365125 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.368819 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.370847 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.407025 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/882e619d-b5ca-4455-abcf-04ee85a45001-utilities\") pod \"redhat-marketplace-5cpfp\" (UID: \"882e619d-b5ca-4455-abcf-04ee85a45001\") " pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.407107 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/882e619d-b5ca-4455-abcf-04ee85a45001-catalog-content\") pod \"redhat-marketplace-5cpfp\" (UID: \"882e619d-b5ca-4455-abcf-04ee85a45001\") " pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.407150 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9mlx\" (UniqueName: \"kubernetes.io/projected/882e619d-b5ca-4455-abcf-04ee85a45001-kube-api-access-g9mlx\") pod \"redhat-marketplace-5cpfp\" (UID: \"882e619d-b5ca-4455-abcf-04ee85a45001\") " pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.427089 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.458010 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" event={"ID":"a6da7528-9b01-40b4-88d3-e78a3c178300","Type":"ContainerStarted","Data":"839ccfda6aed59bced160eea597855bf4711831b79485a05de7c9d5d51274930"} Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.459027 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.463015 4943 generic.go:334] "Generic (PLEG): container finished" podID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" containerID="f87bea5167ff7b26158aabb080b78d71611013b8e5bbe5d1109a4ae8f49c22cf" exitCode=0 Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.463344 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jcsk7" event={"ID":"d22ad643-0a77-49fb-a7f1-54603fe0f82d","Type":"ContainerDied","Data":"f87bea5167ff7b26158aabb080b78d71611013b8e5bbe5d1109a4ae8f49c22cf"} Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.465129 4943 generic.go:334] "Generic (PLEG): container finished" podID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" containerID="7b3a256df2bb19c6f2d22e11c0d4a1f1e5f4a4b0e53e2588c1ec6c403b9bbc6c" exitCode=0 Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.465487 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8cs7" event={"ID":"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565","Type":"ContainerDied","Data":"7b3a256df2bb19c6f2d22e11c0d4a1f1e5f4a4b0e53e2588c1ec6c403b9bbc6c"} Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.466667 4943 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.467617 4943 generic.go:334] "Generic (PLEG): container finished" podID="7462ea89-12b3-4d4a-9757-b2e97b190a53" containerID="b4210dd1392942849874397481182a098959240fcbd65af4618b63b875473fc1" exitCode=0 Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.467673 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkx78" event={"ID":"7462ea89-12b3-4d4a-9757-b2e97b190a53","Type":"ContainerDied","Data":"b4210dd1392942849874397481182a098959240fcbd65af4618b63b875473fc1"} Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.474133 4943 generic.go:334] "Generic (PLEG): container finished" podID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" containerID="5097a1826504d7701f463452245f87b1b01b42e6f863f0fea7f06d282852d7f1" exitCode=0 Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.474224 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7p8s" event={"ID":"0d555ec7-2800-45a7-a1a2-27b1a435b0d4","Type":"ContainerDied","Data":"5097a1826504d7701f463452245f87b1b01b42e6f863f0fea7f06d282852d7f1"} Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.483624 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" podStartSLOduration=137.48360689 podStartE2EDuration="2m17.48360689s" podCreationTimestamp="2025-12-04 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:31.476742702 +0000 UTC m=+160.065718580" watchObservedRunningTime="2025-12-04 10:18:31.48360689 +0000 UTC m=+160.072582758" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.508828 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/882e619d-b5ca-4455-abcf-04ee85a45001-catalog-content\") pod \"redhat-marketplace-5cpfp\" (UID: \"882e619d-b5ca-4455-abcf-04ee85a45001\") " pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.509188 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9mlx\" (UniqueName: \"kubernetes.io/projected/882e619d-b5ca-4455-abcf-04ee85a45001-kube-api-access-g9mlx\") pod \"redhat-marketplace-5cpfp\" (UID: \"882e619d-b5ca-4455-abcf-04ee85a45001\") " pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.509252 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96d31261-5601-4e46-9d92-8172c7a62677-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"96d31261-5601-4e46-9d92-8172c7a62677\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.509350 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/96d31261-5601-4e46-9d92-8172c7a62677-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"96d31261-5601-4e46-9d92-8172c7a62677\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.509419 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/882e619d-b5ca-4455-abcf-04ee85a45001-utilities\") pod \"redhat-marketplace-5cpfp\" (UID: \"882e619d-b5ca-4455-abcf-04ee85a45001\") " pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.509581 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/882e619d-b5ca-4455-abcf-04ee85a45001-catalog-content\") pod \"redhat-marketplace-5cpfp\" (UID: \"882e619d-b5ca-4455-abcf-04ee85a45001\") " pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.509859 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/882e619d-b5ca-4455-abcf-04ee85a45001-utilities\") pod \"redhat-marketplace-5cpfp\" (UID: \"882e619d-b5ca-4455-abcf-04ee85a45001\") " pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.526189 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9mlx\" (UniqueName: \"kubernetes.io/projected/882e619d-b5ca-4455-abcf-04ee85a45001-kube-api-access-g9mlx\") pod \"redhat-marketplace-5cpfp\" (UID: \"882e619d-b5ca-4455-abcf-04ee85a45001\") " pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.568852 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.611136 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/96d31261-5601-4e46-9d92-8172c7a62677-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"96d31261-5601-4e46-9d92-8172c7a62677\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.611523 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96d31261-5601-4e46-9d92-8172c7a62677-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"96d31261-5601-4e46-9d92-8172c7a62677\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.612401 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/96d31261-5601-4e46-9d92-8172c7a62677-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"96d31261-5601-4e46-9d92-8172c7a62677\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.642782 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96d31261-5601-4e46-9d92-8172c7a62677-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"96d31261-5601-4e46-9d92-8172c7a62677\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.702408 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.754187 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9rd8"] Dec 04 10:18:31 crc kubenswrapper[4943]: W1204 10:18:31.781766 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c9ccdf7_acc0_481a_b1de_5172e16ef22c.slice/crio-775938d2763cb2e750e2368545b222f05d565163d9f7ff3a60c338d5eda7bae1 WatchSource:0}: Error finding container 775938d2763cb2e750e2368545b222f05d565163d9f7ff3a60c338d5eda7bae1: Status 404 returned error can't find the container with id 775938d2763cb2e750e2368545b222f05d565163d9f7ff3a60c338d5eda7bae1 Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.835835 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5cpfp"] Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.879868 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gtr9n"] Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.882176 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.885102 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 10:18:31 crc kubenswrapper[4943]: I1204 10:18:31.886273 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gtr9n"] Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.000796 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.016733 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z96bh\" (UniqueName: \"kubernetes.io/projected/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-kube-api-access-z96bh\") pod \"redhat-operators-gtr9n\" (UID: \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\") " pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.016876 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-catalog-content\") pod \"redhat-operators-gtr9n\" (UID: \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\") " pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.016913 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-utilities\") pod \"redhat-operators-gtr9n\" (UID: \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\") " pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.038526 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v2fts"] Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.039620 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.049940 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v2fts"] Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.117828 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-catalog-content\") pod \"redhat-operators-gtr9n\" (UID: \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\") " pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.118160 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-utilities\") pod \"redhat-operators-gtr9n\" (UID: \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\") " pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.118306 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z96bh\" (UniqueName: \"kubernetes.io/projected/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-kube-api-access-z96bh\") pod \"redhat-operators-gtr9n\" (UID: \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\") " pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.118496 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-catalog-content\") pod \"redhat-operators-gtr9n\" (UID: \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\") " pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.118822 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-utilities\") pod \"redhat-operators-gtr9n\" (UID: \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\") " pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.155083 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z96bh\" (UniqueName: \"kubernetes.io/projected/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-kube-api-access-z96bh\") pod \"redhat-operators-gtr9n\" (UID: \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\") " pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.219283 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88878a37-53a8-4edf-8156-ccb1c625ea0d-utilities\") pod \"redhat-operators-v2fts\" (UID: \"88878a37-53a8-4edf-8156-ccb1c625ea0d\") " pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.219376 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88878a37-53a8-4edf-8156-ccb1c625ea0d-catalog-content\") pod \"redhat-operators-v2fts\" (UID: \"88878a37-53a8-4edf-8156-ccb1c625ea0d\") " pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.219617 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj4cn\" (UniqueName: \"kubernetes.io/projected/88878a37-53a8-4edf-8156-ccb1c625ea0d-kube-api-access-bj4cn\") pod \"redhat-operators-v2fts\" (UID: \"88878a37-53a8-4edf-8156-ccb1c625ea0d\") " pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.242932 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:32 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:32 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:32 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.242997 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.264888 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.321103 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88878a37-53a8-4edf-8156-ccb1c625ea0d-catalog-content\") pod \"redhat-operators-v2fts\" (UID: \"88878a37-53a8-4edf-8156-ccb1c625ea0d\") " pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.321270 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj4cn\" (UniqueName: \"kubernetes.io/projected/88878a37-53a8-4edf-8156-ccb1c625ea0d-kube-api-access-bj4cn\") pod \"redhat-operators-v2fts\" (UID: \"88878a37-53a8-4edf-8156-ccb1c625ea0d\") " pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.321356 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88878a37-53a8-4edf-8156-ccb1c625ea0d-utilities\") pod \"redhat-operators-v2fts\" (UID: \"88878a37-53a8-4edf-8156-ccb1c625ea0d\") " pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.322002 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88878a37-53a8-4edf-8156-ccb1c625ea0d-utilities\") pod \"redhat-operators-v2fts\" (UID: \"88878a37-53a8-4edf-8156-ccb1c625ea0d\") " pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.322470 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88878a37-53a8-4edf-8156-ccb1c625ea0d-catalog-content\") pod \"redhat-operators-v2fts\" (UID: \"88878a37-53a8-4edf-8156-ccb1c625ea0d\") " pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.343994 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj4cn\" (UniqueName: \"kubernetes.io/projected/88878a37-53a8-4edf-8156-ccb1c625ea0d-kube-api-access-bj4cn\") pod \"redhat-operators-v2fts\" (UID: \"88878a37-53a8-4edf-8156-ccb1c625ea0d\") " pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.474909 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.489801 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"96d31261-5601-4e46-9d92-8172c7a62677","Type":"ContainerStarted","Data":"d1a1584a050a6f251041619407373855bc97038ca94de2006969145ed38f1618"} Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.493298 4943 generic.go:334] "Generic (PLEG): container finished" podID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" containerID="356985a1da476424d3b644ea0bffaba449408ee868825f7e6734bf11c142f6b0" exitCode=0 Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.493372 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9rd8" event={"ID":"2c9ccdf7-acc0-481a-b1de-5172e16ef22c","Type":"ContainerDied","Data":"356985a1da476424d3b644ea0bffaba449408ee868825f7e6734bf11c142f6b0"} Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.493402 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9rd8" event={"ID":"2c9ccdf7-acc0-481a-b1de-5172e16ef22c","Type":"ContainerStarted","Data":"775938d2763cb2e750e2368545b222f05d565163d9f7ff3a60c338d5eda7bae1"} Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.500186 4943 generic.go:334] "Generic (PLEG): container finished" podID="882e619d-b5ca-4455-abcf-04ee85a45001" containerID="ac4926d1d72f057d1be6cb8d8c34755069890619a95a035abdf35e6502fd5adb" exitCode=0 Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.500483 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cpfp" event={"ID":"882e619d-b5ca-4455-abcf-04ee85a45001","Type":"ContainerDied","Data":"ac4926d1d72f057d1be6cb8d8c34755069890619a95a035abdf35e6502fd5adb"} Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.500567 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cpfp" event={"ID":"882e619d-b5ca-4455-abcf-04ee85a45001","Type":"ContainerStarted","Data":"d0f6d12d6ec29f99e80887b71eea6359ba67c2da2f02178077f12db977e6e6f4"} Dec 04 10:18:32 crc kubenswrapper[4943]: I1204 10:18:32.630482 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gtr9n"] Dec 04 10:18:32 crc kubenswrapper[4943]: W1204 10:18:32.642667 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a3b7592_0bdb_4655_87a3_86ecacc6a0fa.slice/crio-8e5d4f53edddcb2cc65345650130a916767d1b7aa386111ef26c5225aa7ea1af WatchSource:0}: Error finding container 8e5d4f53edddcb2cc65345650130a916767d1b7aa386111ef26c5225aa7ea1af: Status 404 returned error can't find the container with id 8e5d4f53edddcb2cc65345650130a916767d1b7aa386111ef26c5225aa7ea1af Dec 04 10:18:33 crc kubenswrapper[4943]: I1204 10:18:33.030368 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v2fts"] Dec 04 10:18:33 crc kubenswrapper[4943]: W1204 10:18:33.043793 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88878a37_53a8_4edf_8156_ccb1c625ea0d.slice/crio-0652ffb167a3ede5e7c081177adf12fda44c3f96cbd63eb951faa15b44bd646b WatchSource:0}: Error finding container 0652ffb167a3ede5e7c081177adf12fda44c3f96cbd63eb951faa15b44bd646b: Status 404 returned error can't find the container with id 0652ffb167a3ede5e7c081177adf12fda44c3f96cbd63eb951faa15b44bd646b Dec 04 10:18:33 crc kubenswrapper[4943]: I1204 10:18:33.243248 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:33 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:33 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:33 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:33 crc kubenswrapper[4943]: I1204 10:18:33.243325 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:33 crc kubenswrapper[4943]: I1204 10:18:33.559368 4943 generic.go:334] "Generic (PLEG): container finished" podID="88878a37-53a8-4edf-8156-ccb1c625ea0d" containerID="7f4f9a4c8e80138a48303aada99e793b4ef7f547cac69c30c5aed322b513dd4d" exitCode=0 Dec 04 10:18:33 crc kubenswrapper[4943]: I1204 10:18:33.559446 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2fts" event={"ID":"88878a37-53a8-4edf-8156-ccb1c625ea0d","Type":"ContainerDied","Data":"7f4f9a4c8e80138a48303aada99e793b4ef7f547cac69c30c5aed322b513dd4d"} Dec 04 10:18:33 crc kubenswrapper[4943]: I1204 10:18:33.559494 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2fts" event={"ID":"88878a37-53a8-4edf-8156-ccb1c625ea0d","Type":"ContainerStarted","Data":"0652ffb167a3ede5e7c081177adf12fda44c3f96cbd63eb951faa15b44bd646b"} Dec 04 10:18:33 crc kubenswrapper[4943]: I1204 10:18:33.563512 4943 generic.go:334] "Generic (PLEG): container finished" podID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" containerID="98b4197117fbc4df73e705d92977317e05478527b4efefc0e9e63f3f26fdaa47" exitCode=0 Dec 04 10:18:33 crc kubenswrapper[4943]: I1204 10:18:33.563571 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtr9n" event={"ID":"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa","Type":"ContainerDied","Data":"98b4197117fbc4df73e705d92977317e05478527b4efefc0e9e63f3f26fdaa47"} Dec 04 10:18:33 crc kubenswrapper[4943]: I1204 10:18:33.563593 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtr9n" event={"ID":"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa","Type":"ContainerStarted","Data":"8e5d4f53edddcb2cc65345650130a916767d1b7aa386111ef26c5225aa7ea1af"} Dec 04 10:18:33 crc kubenswrapper[4943]: I1204 10:18:33.571222 4943 generic.go:334] "Generic (PLEG): container finished" podID="96d31261-5601-4e46-9d92-8172c7a62677" containerID="2b89c1ca5f6af3aaa51f0090cba0e7a87e9fdcc43d72a84dfa519dba0c05ea9e" exitCode=0 Dec 04 10:18:33 crc kubenswrapper[4943]: I1204 10:18:33.571927 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"96d31261-5601-4e46-9d92-8172c7a62677","Type":"ContainerDied","Data":"2b89c1ca5f6af3aaa51f0090cba0e7a87e9fdcc43d72a84dfa519dba0c05ea9e"} Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.157253 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.158274 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.161701 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.161771 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.174109 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.229638 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/def0a573-07c6-416c-a45d-fc4f469e2330-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"def0a573-07c6-416c-a45d-fc4f469e2330\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.229810 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/def0a573-07c6-416c-a45d-fc4f469e2330-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"def0a573-07c6-416c-a45d-fc4f469e2330\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.245584 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:34 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:34 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:34 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.246089 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.332376 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/def0a573-07c6-416c-a45d-fc4f469e2330-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"def0a573-07c6-416c-a45d-fc4f469e2330\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.332463 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/def0a573-07c6-416c-a45d-fc4f469e2330-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"def0a573-07c6-416c-a45d-fc4f469e2330\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.332610 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/def0a573-07c6-416c-a45d-fc4f469e2330-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"def0a573-07c6-416c-a45d-fc4f469e2330\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.359128 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/def0a573-07c6-416c-a45d-fc4f469e2330-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"def0a573-07c6-416c-a45d-fc4f469e2330\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.495512 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.678698 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:34 crc kubenswrapper[4943]: I1204 10:18:34.684960 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-5bgm4" Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.246771 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:35 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:35 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:35 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.246856 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.269046 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-fs4kj" Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.352623 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.361765 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.511152 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96d31261-5601-4e46-9d92-8172c7a62677-kube-api-access\") pod \"96d31261-5601-4e46-9d92-8172c7a62677\" (UID: \"96d31261-5601-4e46-9d92-8172c7a62677\") " Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.511236 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/96d31261-5601-4e46-9d92-8172c7a62677-kubelet-dir\") pod \"96d31261-5601-4e46-9d92-8172c7a62677\" (UID: \"96d31261-5601-4e46-9d92-8172c7a62677\") " Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.511575 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/96d31261-5601-4e46-9d92-8172c7a62677-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "96d31261-5601-4e46-9d92-8172c7a62677" (UID: "96d31261-5601-4e46-9d92-8172c7a62677"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.526621 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96d31261-5601-4e46-9d92-8172c7a62677-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "96d31261-5601-4e46-9d92-8172c7a62677" (UID: "96d31261-5601-4e46-9d92-8172c7a62677"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.612727 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/96d31261-5601-4e46-9d92-8172c7a62677-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.612775 4943 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/96d31261-5601-4e46-9d92-8172c7a62677-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.667823 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.667973 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"96d31261-5601-4e46-9d92-8172c7a62677","Type":"ContainerDied","Data":"d1a1584a050a6f251041619407373855bc97038ca94de2006969145ed38f1618"} Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.668031 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1a1584a050a6f251041619407373855bc97038ca94de2006969145ed38f1618" Dec 04 10:18:35 crc kubenswrapper[4943]: I1204 10:18:35.684844 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"def0a573-07c6-416c-a45d-fc4f469e2330","Type":"ContainerStarted","Data":"b55b1e7ec47bba867fe98322da49b5bfa807415cb527a1b831aa1d63ae70bd90"} Dec 04 10:18:36 crc kubenswrapper[4943]: I1204 10:18:36.243919 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:36 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:36 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:36 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:36 crc kubenswrapper[4943]: I1204 10:18:36.243984 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:36 crc kubenswrapper[4943]: I1204 10:18:36.336146 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:18:36 crc kubenswrapper[4943]: I1204 10:18:36.346649 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb85a8e0-6c90-47cb-a1b5-9ecd5244c710-metrics-certs\") pod \"network-metrics-daemon-hw5pj\" (UID: \"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710\") " pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:18:36 crc kubenswrapper[4943]: I1204 10:18:36.482729 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hw5pj" Dec 04 10:18:37 crc kubenswrapper[4943]: I1204 10:18:37.064194 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hw5pj"] Dec 04 10:18:37 crc kubenswrapper[4943]: W1204 10:18:37.083583 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb85a8e0_6c90_47cb_a1b5_9ecd5244c710.slice/crio-10c008ace808bf953f801e5449581ff4a61097e4ffd10109470767508ecca88c WatchSource:0}: Error finding container 10c008ace808bf953f801e5449581ff4a61097e4ffd10109470767508ecca88c: Status 404 returned error can't find the container with id 10c008ace808bf953f801e5449581ff4a61097e4ffd10109470767508ecca88c Dec 04 10:18:37 crc kubenswrapper[4943]: I1204 10:18:37.243669 4943 patch_prober.go:28] interesting pod/router-default-5444994796-zkjfk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 10:18:37 crc kubenswrapper[4943]: [-]has-synced failed: reason withheld Dec 04 10:18:37 crc kubenswrapper[4943]: [+]process-running ok Dec 04 10:18:37 crc kubenswrapper[4943]: healthz check failed Dec 04 10:18:37 crc kubenswrapper[4943]: I1204 10:18:37.243744 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zkjfk" podUID="cb5f8eaa-9da1-4a88-b277-b8f76e86123d" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 10:18:37 crc kubenswrapper[4943]: I1204 10:18:37.724817 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" event={"ID":"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710","Type":"ContainerStarted","Data":"10c008ace808bf953f801e5449581ff4a61097e4ffd10109470767508ecca88c"} Dec 04 10:18:38 crc kubenswrapper[4943]: I1204 10:18:38.247958 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:38 crc kubenswrapper[4943]: I1204 10:18:38.252413 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-zkjfk" Dec 04 10:18:38 crc kubenswrapper[4943]: I1204 10:18:38.768116 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"def0a573-07c6-416c-a45d-fc4f469e2330","Type":"ContainerStarted","Data":"fd67a183138ce6cb763bba680f72c2290a8581823fadb55d33768998c6d1ccc4"} Dec 04 10:18:38 crc kubenswrapper[4943]: I1204 10:18:38.790509 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.790485167 podStartE2EDuration="4.790485167s" podCreationTimestamp="2025-12-04 10:18:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:38.788392406 +0000 UTC m=+167.377368294" watchObservedRunningTime="2025-12-04 10:18:38.790485167 +0000 UTC m=+167.379461035" Dec 04 10:18:39 crc kubenswrapper[4943]: I1204 10:18:39.785850 4943 generic.go:334] "Generic (PLEG): container finished" podID="def0a573-07c6-416c-a45d-fc4f469e2330" containerID="fd67a183138ce6cb763bba680f72c2290a8581823fadb55d33768998c6d1ccc4" exitCode=0 Dec 04 10:18:39 crc kubenswrapper[4943]: I1204 10:18:39.785962 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"def0a573-07c6-416c-a45d-fc4f469e2330","Type":"ContainerDied","Data":"fd67a183138ce6cb763bba680f72c2290a8581823fadb55d33768998c6d1ccc4"} Dec 04 10:18:39 crc kubenswrapper[4943]: I1204 10:18:39.791220 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" event={"ID":"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710","Type":"ContainerStarted","Data":"7bfa7e798ff6523e9954bdf1b2db61db8c993f976e072af0c289c345ef9cd718"} Dec 04 10:18:39 crc kubenswrapper[4943]: I1204 10:18:39.872042 4943 patch_prober.go:28] interesting pod/downloads-7954f5f757-grw5p container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 04 10:18:39 crc kubenswrapper[4943]: I1204 10:18:39.872118 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-grw5p" podUID="1abfb759-04f9-4885-b150-82355f8cf866" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 04 10:18:39 crc kubenswrapper[4943]: I1204 10:18:39.872156 4943 patch_prober.go:28] interesting pod/downloads-7954f5f757-grw5p container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 04 10:18:39 crc kubenswrapper[4943]: I1204 10:18:39.872311 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-grw5p" podUID="1abfb759-04f9-4885-b150-82355f8cf866" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 04 10:18:39 crc kubenswrapper[4943]: I1204 10:18:39.980478 4943 patch_prober.go:28] interesting pod/console-f9d7485db-kdt7l container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 04 10:18:39 crc kubenswrapper[4943]: I1204 10:18:39.980554 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-kdt7l" podUID="7848703f-6d8c-4915-8f42-e6222f2a889a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 04 10:18:40 crc kubenswrapper[4943]: I1204 10:18:40.814393 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hw5pj" event={"ID":"eb85a8e0-6c90-47cb-a1b5-9ecd5244c710","Type":"ContainerStarted","Data":"897986d6a9992577218fa64f66b96acd4e6d5cd93fba8ed43b4029c137bd42b1"} Dec 04 10:18:40 crc kubenswrapper[4943]: I1204 10:18:40.837180 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-hw5pj" podStartSLOduration=147.837151942 podStartE2EDuration="2m27.837151942s" podCreationTimestamp="2025-12-04 10:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:18:40.83581919 +0000 UTC m=+169.424795058" watchObservedRunningTime="2025-12-04 10:18:40.837151942 +0000 UTC m=+169.426127810" Dec 04 10:18:48 crc kubenswrapper[4943]: I1204 10:18:48.629088 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:18:48 crc kubenswrapper[4943]: I1204 10:18:48.629713 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:18:48 crc kubenswrapper[4943]: I1204 10:18:48.797325 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.059400 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.071355 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"def0a573-07c6-416c-a45d-fc4f469e2330","Type":"ContainerDied","Data":"b55b1e7ec47bba867fe98322da49b5bfa807415cb527a1b831aa1d63ae70bd90"} Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.071532 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b55b1e7ec47bba867fe98322da49b5bfa807415cb527a1b831aa1d63ae70bd90" Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.071433 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.181597 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/def0a573-07c6-416c-a45d-fc4f469e2330-kube-api-access\") pod \"def0a573-07c6-416c-a45d-fc4f469e2330\" (UID: \"def0a573-07c6-416c-a45d-fc4f469e2330\") " Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.181705 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/def0a573-07c6-416c-a45d-fc4f469e2330-kubelet-dir\") pod \"def0a573-07c6-416c-a45d-fc4f469e2330\" (UID: \"def0a573-07c6-416c-a45d-fc4f469e2330\") " Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.181933 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/def0a573-07c6-416c-a45d-fc4f469e2330-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "def0a573-07c6-416c-a45d-fc4f469e2330" (UID: "def0a573-07c6-416c-a45d-fc4f469e2330"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.188074 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/def0a573-07c6-416c-a45d-fc4f469e2330-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "def0a573-07c6-416c-a45d-fc4f469e2330" (UID: "def0a573-07c6-416c-a45d-fc4f469e2330"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.282927 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/def0a573-07c6-416c-a45d-fc4f469e2330-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.282975 4943 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/def0a573-07c6-416c-a45d-fc4f469e2330-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.876032 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-grw5p" Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.985024 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:49 crc kubenswrapper[4943]: I1204 10:18:49.990910 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-kdt7l" Dec 04 10:18:59 crc kubenswrapper[4943]: I1204 10:18:59.463298 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ws8hv" Dec 04 10:19:01 crc kubenswrapper[4943]: I1204 10:19:01.641567 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 10:19:09 crc kubenswrapper[4943]: E1204 10:19:09.796952 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 04 10:19:09 crc kubenswrapper[4943]: E1204 10:19:09.798278 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8wzkd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-b8cs7_openshift-marketplace(8c03ef2d-4f6e-48c5-b6f9-825fc02b8565): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 10:19:09 crc kubenswrapper[4943]: E1204 10:19:09.799539 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-b8cs7" podUID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" Dec 04 10:19:10 crc kubenswrapper[4943]: I1204 10:19:10.951983 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 10:19:10 crc kubenswrapper[4943]: E1204 10:19:10.953510 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def0a573-07c6-416c-a45d-fc4f469e2330" containerName="pruner" Dec 04 10:19:10 crc kubenswrapper[4943]: I1204 10:19:10.953546 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="def0a573-07c6-416c-a45d-fc4f469e2330" containerName="pruner" Dec 04 10:19:10 crc kubenswrapper[4943]: E1204 10:19:10.953580 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96d31261-5601-4e46-9d92-8172c7a62677" containerName="pruner" Dec 04 10:19:10 crc kubenswrapper[4943]: I1204 10:19:10.953588 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="96d31261-5601-4e46-9d92-8172c7a62677" containerName="pruner" Dec 04 10:19:10 crc kubenswrapper[4943]: I1204 10:19:10.955956 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="96d31261-5601-4e46-9d92-8172c7a62677" containerName="pruner" Dec 04 10:19:10 crc kubenswrapper[4943]: I1204 10:19:10.957377 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="def0a573-07c6-416c-a45d-fc4f469e2330" containerName="pruner" Dec 04 10:19:10 crc kubenswrapper[4943]: I1204 10:19:10.958011 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 10:19:10 crc kubenswrapper[4943]: I1204 10:19:10.958120 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 10:19:10 crc kubenswrapper[4943]: I1204 10:19:10.960327 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 10:19:10 crc kubenswrapper[4943]: I1204 10:19:10.960496 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 10:19:11 crc kubenswrapper[4943]: I1204 10:19:11.075958 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e67b367b-a25a-452b-9e9b-3af2cdaef856-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e67b367b-a25a-452b-9e9b-3af2cdaef856\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 10:19:11 crc kubenswrapper[4943]: I1204 10:19:11.076018 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e67b367b-a25a-452b-9e9b-3af2cdaef856-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e67b367b-a25a-452b-9e9b-3af2cdaef856\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 10:19:11 crc kubenswrapper[4943]: I1204 10:19:11.177663 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e67b367b-a25a-452b-9e9b-3af2cdaef856-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e67b367b-a25a-452b-9e9b-3af2cdaef856\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 10:19:11 crc kubenswrapper[4943]: I1204 10:19:11.177725 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e67b367b-a25a-452b-9e9b-3af2cdaef856-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e67b367b-a25a-452b-9e9b-3af2cdaef856\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 10:19:11 crc kubenswrapper[4943]: I1204 10:19:11.177855 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e67b367b-a25a-452b-9e9b-3af2cdaef856-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e67b367b-a25a-452b-9e9b-3af2cdaef856\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 10:19:11 crc kubenswrapper[4943]: I1204 10:19:11.197073 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e67b367b-a25a-452b-9e9b-3af2cdaef856-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e67b367b-a25a-452b-9e9b-3af2cdaef856\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 10:19:11 crc kubenswrapper[4943]: I1204 10:19:11.285290 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 10:19:11 crc kubenswrapper[4943]: E1204 10:19:11.803548 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-b8cs7" podUID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" Dec 04 10:19:11 crc kubenswrapper[4943]: E1204 10:19:11.949059 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 04 10:19:11 crc kubenswrapper[4943]: E1204 10:19:11.949505 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rfjpc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-jcsk7_openshift-marketplace(d22ad643-0a77-49fb-a7f1-54603fe0f82d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 10:19:11 crc kubenswrapper[4943]: E1204 10:19:11.951579 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-jcsk7" podUID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" Dec 04 10:19:15 crc kubenswrapper[4943]: I1204 10:19:15.917935 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 10:19:15 crc kubenswrapper[4943]: I1204 10:19:15.918792 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:19:15 crc kubenswrapper[4943]: I1204 10:19:15.935425 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 10:19:16 crc kubenswrapper[4943]: I1204 10:19:16.047252 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1683641e-8c40-4be2-b18b-af072d514b99-kubelet-dir\") pod \"installer-9-crc\" (UID: \"1683641e-8c40-4be2-b18b-af072d514b99\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:19:16 crc kubenswrapper[4943]: I1204 10:19:16.047313 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1683641e-8c40-4be2-b18b-af072d514b99-kube-api-access\") pod \"installer-9-crc\" (UID: \"1683641e-8c40-4be2-b18b-af072d514b99\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:19:16 crc kubenswrapper[4943]: I1204 10:19:16.047339 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1683641e-8c40-4be2-b18b-af072d514b99-var-lock\") pod \"installer-9-crc\" (UID: \"1683641e-8c40-4be2-b18b-af072d514b99\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:19:16 crc kubenswrapper[4943]: I1204 10:19:16.149124 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1683641e-8c40-4be2-b18b-af072d514b99-kubelet-dir\") pod \"installer-9-crc\" (UID: \"1683641e-8c40-4be2-b18b-af072d514b99\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:19:16 crc kubenswrapper[4943]: I1204 10:19:16.149237 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1683641e-8c40-4be2-b18b-af072d514b99-var-lock\") pod \"installer-9-crc\" (UID: \"1683641e-8c40-4be2-b18b-af072d514b99\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:19:16 crc kubenswrapper[4943]: I1204 10:19:16.149264 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1683641e-8c40-4be2-b18b-af072d514b99-kube-api-access\") pod \"installer-9-crc\" (UID: \"1683641e-8c40-4be2-b18b-af072d514b99\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:19:16 crc kubenswrapper[4943]: I1204 10:19:16.149314 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1683641e-8c40-4be2-b18b-af072d514b99-kubelet-dir\") pod \"installer-9-crc\" (UID: \"1683641e-8c40-4be2-b18b-af072d514b99\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:19:16 crc kubenswrapper[4943]: I1204 10:19:16.149456 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1683641e-8c40-4be2-b18b-af072d514b99-var-lock\") pod \"installer-9-crc\" (UID: \"1683641e-8c40-4be2-b18b-af072d514b99\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:19:16 crc kubenswrapper[4943]: I1204 10:19:16.171345 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1683641e-8c40-4be2-b18b-af072d514b99-kube-api-access\") pod \"installer-9-crc\" (UID: \"1683641e-8c40-4be2-b18b-af072d514b99\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:19:16 crc kubenswrapper[4943]: I1204 10:19:16.235327 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:19:18 crc kubenswrapper[4943]: I1204 10:19:18.629365 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:19:18 crc kubenswrapper[4943]: I1204 10:19:18.629772 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:19:18 crc kubenswrapper[4943]: I1204 10:19:18.629846 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:19:18 crc kubenswrapper[4943]: I1204 10:19:18.630624 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58"} pod="openshift-machine-config-operator/machine-config-daemon-kswzd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 10:19:18 crc kubenswrapper[4943]: I1204 10:19:18.630977 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" containerID="cri-o://9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58" gracePeriod=600 Dec 04 10:19:18 crc kubenswrapper[4943]: E1204 10:19:18.899307 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-jcsk7" podUID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" Dec 04 10:19:18 crc kubenswrapper[4943]: E1204 10:19:18.980280 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 04 10:19:18 crc kubenswrapper[4943]: E1204 10:19:18.980467 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bj4cn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-v2fts_openshift-marketplace(88878a37-53a8-4edf-8156-ccb1c625ea0d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 10:19:18 crc kubenswrapper[4943]: E1204 10:19:18.981759 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-v2fts" podUID="88878a37-53a8-4edf-8156-ccb1c625ea0d" Dec 04 10:19:19 crc kubenswrapper[4943]: E1204 10:19:19.011870 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 04 10:19:19 crc kubenswrapper[4943]: E1204 10:19:19.012059 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-88vlp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lkx78_openshift-marketplace(7462ea89-12b3-4d4a-9757-b2e97b190a53): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 10:19:19 crc kubenswrapper[4943]: E1204 10:19:19.013444 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-lkx78" podUID="7462ea89-12b3-4d4a-9757-b2e97b190a53" Dec 04 10:19:19 crc kubenswrapper[4943]: E1204 10:19:19.032807 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 04 10:19:19 crc kubenswrapper[4943]: E1204 10:19:19.032960 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gwhw9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-l7p8s_openshift-marketplace(0d555ec7-2800-45a7-a1a2-27b1a435b0d4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 10:19:19 crc kubenswrapper[4943]: E1204 10:19:19.034126 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-l7p8s" podUID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" Dec 04 10:19:19 crc kubenswrapper[4943]: I1204 10:19:19.401615 4943 generic.go:334] "Generic (PLEG): container finished" podID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerID="9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58" exitCode=0 Dec 04 10:19:19 crc kubenswrapper[4943]: I1204 10:19:19.401733 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerDied","Data":"9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58"} Dec 04 10:19:19 crc kubenswrapper[4943]: E1204 10:19:19.997161 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lkx78" podUID="7462ea89-12b3-4d4a-9757-b2e97b190a53" Dec 04 10:19:19 crc kubenswrapper[4943]: E1204 10:19:19.997291 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-l7p8s" podUID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" Dec 04 10:19:19 crc kubenswrapper[4943]: E1204 10:19:19.997362 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-v2fts" podUID="88878a37-53a8-4edf-8156-ccb1c625ea0d" Dec 04 10:19:20 crc kubenswrapper[4943]: E1204 10:19:20.094926 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 10:19:20 crc kubenswrapper[4943]: E1204 10:19:20.095300 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-99lx2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-q9rd8_openshift-marketplace(2c9ccdf7-acc0-481a-b1de-5172e16ef22c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 10:19:20 crc kubenswrapper[4943]: E1204 10:19:20.096471 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-q9rd8" podUID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" Dec 04 10:19:20 crc kubenswrapper[4943]: E1204 10:19:20.096552 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 10:19:20 crc kubenswrapper[4943]: E1204 10:19:20.096689 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g9mlx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5cpfp_openshift-marketplace(882e619d-b5ca-4455-abcf-04ee85a45001): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 10:19:20 crc kubenswrapper[4943]: E1204 10:19:20.097854 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5cpfp" podUID="882e619d-b5ca-4455-abcf-04ee85a45001" Dec 04 10:19:20 crc kubenswrapper[4943]: E1204 10:19:20.113339 4943 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 04 10:19:20 crc kubenswrapper[4943]: E1204 10:19:20.113843 4943 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z96bh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-gtr9n_openshift-marketplace(3a3b7592-0bdb-4655-87a3-86ecacc6a0fa): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 10:19:20 crc kubenswrapper[4943]: E1204 10:19:20.114986 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-gtr9n" podUID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" Dec 04 10:19:20 crc kubenswrapper[4943]: E1204 10:19:20.409055 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-5cpfp" podUID="882e619d-b5ca-4455-abcf-04ee85a45001" Dec 04 10:19:20 crc kubenswrapper[4943]: E1204 10:19:20.409411 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-gtr9n" podUID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" Dec 04 10:19:20 crc kubenswrapper[4943]: E1204 10:19:20.409420 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-q9rd8" podUID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" Dec 04 10:19:20 crc kubenswrapper[4943]: I1204 10:19:20.475769 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 10:19:20 crc kubenswrapper[4943]: W1204 10:19:20.489915 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod1683641e_8c40_4be2_b18b_af072d514b99.slice/crio-dffaeb658bbcd012870dbfb4e007a9301a7590570dd20e740742a12d0d68bed9 WatchSource:0}: Error finding container dffaeb658bbcd012870dbfb4e007a9301a7590570dd20e740742a12d0d68bed9: Status 404 returned error can't find the container with id dffaeb658bbcd012870dbfb4e007a9301a7590570dd20e740742a12d0d68bed9 Dec 04 10:19:20 crc kubenswrapper[4943]: I1204 10:19:20.546087 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 10:19:20 crc kubenswrapper[4943]: W1204 10:19:20.549731 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode67b367b_a25a_452b_9e9b_3af2cdaef856.slice/crio-33dbeff278cc52b3717af24d4db23934474a8c2a553ba3e96c99001ef3432a3f WatchSource:0}: Error finding container 33dbeff278cc52b3717af24d4db23934474a8c2a553ba3e96c99001ef3432a3f: Status 404 returned error can't find the container with id 33dbeff278cc52b3717af24d4db23934474a8c2a553ba3e96c99001ef3432a3f Dec 04 10:19:21 crc kubenswrapper[4943]: I1204 10:19:21.416083 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e67b367b-a25a-452b-9e9b-3af2cdaef856","Type":"ContainerStarted","Data":"33dbeff278cc52b3717af24d4db23934474a8c2a553ba3e96c99001ef3432a3f"} Dec 04 10:19:21 crc kubenswrapper[4943]: I1204 10:19:21.417332 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1683641e-8c40-4be2-b18b-af072d514b99","Type":"ContainerStarted","Data":"dffaeb658bbcd012870dbfb4e007a9301a7590570dd20e740742a12d0d68bed9"} Dec 04 10:19:22 crc kubenswrapper[4943]: I1204 10:19:22.427228 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1683641e-8c40-4be2-b18b-af072d514b99","Type":"ContainerStarted","Data":"3da7a9db0af1031e9b446179144f5cf2cdde42a164f0c01ab6baf16c7e4bcafb"} Dec 04 10:19:22 crc kubenswrapper[4943]: I1204 10:19:22.431859 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerStarted","Data":"98a02a23954dbf78c645fbdf023f59c0401bc3eb6dc01df9396ba4979fbe8194"} Dec 04 10:19:22 crc kubenswrapper[4943]: I1204 10:19:22.434063 4943 generic.go:334] "Generic (PLEG): container finished" podID="e67b367b-a25a-452b-9e9b-3af2cdaef856" containerID="4c5ff8f7f9740fa065878943370d4050ca80ec0f11ea36a797328f51b8f4c1b4" exitCode=0 Dec 04 10:19:22 crc kubenswrapper[4943]: I1204 10:19:22.434114 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e67b367b-a25a-452b-9e9b-3af2cdaef856","Type":"ContainerDied","Data":"4c5ff8f7f9740fa065878943370d4050ca80ec0f11ea36a797328f51b8f4c1b4"} Dec 04 10:19:22 crc kubenswrapper[4943]: I1204 10:19:22.448265 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=7.448190144 podStartE2EDuration="7.448190144s" podCreationTimestamp="2025-12-04 10:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:19:22.442933926 +0000 UTC m=+211.031909794" watchObservedRunningTime="2025-12-04 10:19:22.448190144 +0000 UTC m=+211.037166012" Dec 04 10:19:24 crc kubenswrapper[4943]: I1204 10:19:24.050912 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 10:19:24 crc kubenswrapper[4943]: I1204 10:19:24.156314 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e67b367b-a25a-452b-9e9b-3af2cdaef856-kube-api-access\") pod \"e67b367b-a25a-452b-9e9b-3af2cdaef856\" (UID: \"e67b367b-a25a-452b-9e9b-3af2cdaef856\") " Dec 04 10:19:24 crc kubenswrapper[4943]: I1204 10:19:24.156538 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e67b367b-a25a-452b-9e9b-3af2cdaef856-kubelet-dir\") pod \"e67b367b-a25a-452b-9e9b-3af2cdaef856\" (UID: \"e67b367b-a25a-452b-9e9b-3af2cdaef856\") " Dec 04 10:19:24 crc kubenswrapper[4943]: I1204 10:19:24.156674 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e67b367b-a25a-452b-9e9b-3af2cdaef856-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e67b367b-a25a-452b-9e9b-3af2cdaef856" (UID: "e67b367b-a25a-452b-9e9b-3af2cdaef856"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:19:24 crc kubenswrapper[4943]: I1204 10:19:24.156840 4943 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e67b367b-a25a-452b-9e9b-3af2cdaef856-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:24 crc kubenswrapper[4943]: I1204 10:19:24.162291 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e67b367b-a25a-452b-9e9b-3af2cdaef856-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e67b367b-a25a-452b-9e9b-3af2cdaef856" (UID: "e67b367b-a25a-452b-9e9b-3af2cdaef856"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:19:24 crc kubenswrapper[4943]: I1204 10:19:24.257825 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e67b367b-a25a-452b-9e9b-3af2cdaef856-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:24 crc kubenswrapper[4943]: I1204 10:19:24.448186 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e67b367b-a25a-452b-9e9b-3af2cdaef856","Type":"ContainerDied","Data":"33dbeff278cc52b3717af24d4db23934474a8c2a553ba3e96c99001ef3432a3f"} Dec 04 10:19:24 crc kubenswrapper[4943]: I1204 10:19:24.448543 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33dbeff278cc52b3717af24d4db23934474a8c2a553ba3e96c99001ef3432a3f" Dec 04 10:19:24 crc kubenswrapper[4943]: I1204 10:19:24.448379 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 10:19:26 crc kubenswrapper[4943]: I1204 10:19:26.460759 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8cs7" event={"ID":"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565","Type":"ContainerStarted","Data":"c969ab58fdcc2f9af8442794c530b6f6c0394ced30f42cad776b71daf4e135ce"} Dec 04 10:19:27 crc kubenswrapper[4943]: I1204 10:19:27.470765 4943 generic.go:334] "Generic (PLEG): container finished" podID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" containerID="c969ab58fdcc2f9af8442794c530b6f6c0394ced30f42cad776b71daf4e135ce" exitCode=0 Dec 04 10:19:27 crc kubenswrapper[4943]: I1204 10:19:27.470835 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8cs7" event={"ID":"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565","Type":"ContainerDied","Data":"c969ab58fdcc2f9af8442794c530b6f6c0394ced30f42cad776b71daf4e135ce"} Dec 04 10:19:28 crc kubenswrapper[4943]: I1204 10:19:28.479783 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8cs7" event={"ID":"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565","Type":"ContainerStarted","Data":"e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7"} Dec 04 10:19:28 crc kubenswrapper[4943]: I1204 10:19:28.502138 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b8cs7" podStartSLOduration=3.044208238 podStartE2EDuration="59.502094209s" podCreationTimestamp="2025-12-04 10:18:29 +0000 UTC" firstStartedPulling="2025-12-04 10:18:31.466413669 +0000 UTC m=+160.055389537" lastFinishedPulling="2025-12-04 10:19:27.92429964 +0000 UTC m=+216.513275508" observedRunningTime="2025-12-04 10:19:28.498775127 +0000 UTC m=+217.087751015" watchObservedRunningTime="2025-12-04 10:19:28.502094209 +0000 UTC m=+217.091070077" Dec 04 10:19:29 crc kubenswrapper[4943]: I1204 10:19:29.723646 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:19:29 crc kubenswrapper[4943]: I1204 10:19:29.723997 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:19:30 crc kubenswrapper[4943]: I1204 10:19:30.235914 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:19:33 crc kubenswrapper[4943]: I1204 10:19:33.611142 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2fts" event={"ID":"88878a37-53a8-4edf-8156-ccb1c625ea0d","Type":"ContainerStarted","Data":"0c4b6b95cfea51fac5c1a00f12711dae33c257ea8138389d66d07fa2dd9cdd2f"} Dec 04 10:19:33 crc kubenswrapper[4943]: I1204 10:19:33.613262 4943 generic.go:334] "Generic (PLEG): container finished" podID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" containerID="12da88b3bc918fe777d34adea50a0b4eb3b7b7379221e26f65414fc9b5e59071" exitCode=0 Dec 04 10:19:33 crc kubenswrapper[4943]: I1204 10:19:33.613290 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jcsk7" event={"ID":"d22ad643-0a77-49fb-a7f1-54603fe0f82d","Type":"ContainerDied","Data":"12da88b3bc918fe777d34adea50a0b4eb3b7b7379221e26f65414fc9b5e59071"} Dec 04 10:19:34 crc kubenswrapper[4943]: I1204 10:19:34.627558 4943 generic.go:334] "Generic (PLEG): container finished" podID="88878a37-53a8-4edf-8156-ccb1c625ea0d" containerID="0c4b6b95cfea51fac5c1a00f12711dae33c257ea8138389d66d07fa2dd9cdd2f" exitCode=0 Dec 04 10:19:34 crc kubenswrapper[4943]: I1204 10:19:34.627663 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2fts" event={"ID":"88878a37-53a8-4edf-8156-ccb1c625ea0d","Type":"ContainerDied","Data":"0c4b6b95cfea51fac5c1a00f12711dae33c257ea8138389d66d07fa2dd9cdd2f"} Dec 04 10:19:34 crc kubenswrapper[4943]: I1204 10:19:34.634341 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jcsk7" event={"ID":"d22ad643-0a77-49fb-a7f1-54603fe0f82d","Type":"ContainerStarted","Data":"284bb10b585888ced45051d724242bdb9424f39fd99c729021086fbb344cf1a4"} Dec 04 10:19:34 crc kubenswrapper[4943]: I1204 10:19:34.721735 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jcsk7" podStartSLOduration=4.061776231 podStartE2EDuration="1m6.721708231s" podCreationTimestamp="2025-12-04 10:18:28 +0000 UTC" firstStartedPulling="2025-12-04 10:18:31.46605397 +0000 UTC m=+160.055029838" lastFinishedPulling="2025-12-04 10:19:34.12598596 +0000 UTC m=+222.714961838" observedRunningTime="2025-12-04 10:19:34.699859034 +0000 UTC m=+223.288834902" watchObservedRunningTime="2025-12-04 10:19:34.721708231 +0000 UTC m=+223.310684099" Dec 04 10:19:35 crc kubenswrapper[4943]: I1204 10:19:35.641905 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkx78" event={"ID":"7462ea89-12b3-4d4a-9757-b2e97b190a53","Type":"ContainerStarted","Data":"9f1ed6045be67d4f4cee9dd1fcd2ee0bcaf577aa8cf3fafd1a57a8c08ad7abbc"} Dec 04 10:19:35 crc kubenswrapper[4943]: I1204 10:19:35.643945 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cpfp" event={"ID":"882e619d-b5ca-4455-abcf-04ee85a45001","Type":"ContainerStarted","Data":"4bd7da92ef80a5c2b0228d2ffea4e81eb50c8722b16fb866f7d58fee35ee4241"} Dec 04 10:19:35 crc kubenswrapper[4943]: I1204 10:19:35.645620 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7p8s" event={"ID":"0d555ec7-2800-45a7-a1a2-27b1a435b0d4","Type":"ContainerStarted","Data":"b49d16cea7fdbb8aaa63a329bf68f393dd91587a2c440d2875542cb1d80c58c2"} Dec 04 10:19:36 crc kubenswrapper[4943]: I1204 10:19:36.813407 4943 generic.go:334] "Generic (PLEG): container finished" podID="7462ea89-12b3-4d4a-9757-b2e97b190a53" containerID="9f1ed6045be67d4f4cee9dd1fcd2ee0bcaf577aa8cf3fafd1a57a8c08ad7abbc" exitCode=0 Dec 04 10:19:36 crc kubenswrapper[4943]: I1204 10:19:36.813928 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkx78" event={"ID":"7462ea89-12b3-4d4a-9757-b2e97b190a53","Type":"ContainerDied","Data":"9f1ed6045be67d4f4cee9dd1fcd2ee0bcaf577aa8cf3fafd1a57a8c08ad7abbc"} Dec 04 10:19:36 crc kubenswrapper[4943]: I1204 10:19:36.830510 4943 generic.go:334] "Generic (PLEG): container finished" podID="882e619d-b5ca-4455-abcf-04ee85a45001" containerID="4bd7da92ef80a5c2b0228d2ffea4e81eb50c8722b16fb866f7d58fee35ee4241" exitCode=0 Dec 04 10:19:36 crc kubenswrapper[4943]: I1204 10:19:36.830655 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cpfp" event={"ID":"882e619d-b5ca-4455-abcf-04ee85a45001","Type":"ContainerDied","Data":"4bd7da92ef80a5c2b0228d2ffea4e81eb50c8722b16fb866f7d58fee35ee4241"} Dec 04 10:19:36 crc kubenswrapper[4943]: I1204 10:19:36.842532 4943 generic.go:334] "Generic (PLEG): container finished" podID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" containerID="b49d16cea7fdbb8aaa63a329bf68f393dd91587a2c440d2875542cb1d80c58c2" exitCode=0 Dec 04 10:19:36 crc kubenswrapper[4943]: I1204 10:19:36.842588 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7p8s" event={"ID":"0d555ec7-2800-45a7-a1a2-27b1a435b0d4","Type":"ContainerDied","Data":"b49d16cea7fdbb8aaa63a329bf68f393dd91587a2c440d2875542cb1d80c58c2"} Dec 04 10:19:36 crc kubenswrapper[4943]: I1204 10:19:36.848057 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtr9n" event={"ID":"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa","Type":"ContainerStarted","Data":"a826e9bae054356100ca51c7ea5adc95f02d735ac4ba137ebdbf0b4987098016"} Dec 04 10:19:37 crc kubenswrapper[4943]: I1204 10:19:37.909927 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cpfp" event={"ID":"882e619d-b5ca-4455-abcf-04ee85a45001","Type":"ContainerStarted","Data":"5797fb1ce00dfe2ef33e3ec217f72e6ee9473cf5d28d72028fe50dbdc1e92b50"} Dec 04 10:19:37 crc kubenswrapper[4943]: I1204 10:19:37.913999 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7p8s" event={"ID":"0d555ec7-2800-45a7-a1a2-27b1a435b0d4","Type":"ContainerStarted","Data":"5d7a4fb66febdd49aca0d0f99bb524bb59009d29c5081a886304e5ee3a455c6f"} Dec 04 10:19:37 crc kubenswrapper[4943]: I1204 10:19:37.917447 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2fts" event={"ID":"88878a37-53a8-4edf-8156-ccb1c625ea0d","Type":"ContainerStarted","Data":"d5b69a44d50a9d99366606e39180a6bb372692b9f3f87013f6bc16a7dddbdacb"} Dec 04 10:19:37 crc kubenswrapper[4943]: I1204 10:19:37.919870 4943 generic.go:334] "Generic (PLEG): container finished" podID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" containerID="a826e9bae054356100ca51c7ea5adc95f02d735ac4ba137ebdbf0b4987098016" exitCode=0 Dec 04 10:19:37 crc kubenswrapper[4943]: I1204 10:19:37.919925 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtr9n" event={"ID":"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa","Type":"ContainerDied","Data":"a826e9bae054356100ca51c7ea5adc95f02d735ac4ba137ebdbf0b4987098016"} Dec 04 10:19:37 crc kubenswrapper[4943]: I1204 10:19:37.923051 4943 generic.go:334] "Generic (PLEG): container finished" podID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" containerID="7d856df5c5c7a8fc2ad0d6654f66e626ab449b327f45a5e611e60fa0c0b1fb1c" exitCode=0 Dec 04 10:19:37 crc kubenswrapper[4943]: I1204 10:19:37.923099 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9rd8" event={"ID":"2c9ccdf7-acc0-481a-b1de-5172e16ef22c","Type":"ContainerDied","Data":"7d856df5c5c7a8fc2ad0d6654f66e626ab449b327f45a5e611e60fa0c0b1fb1c"} Dec 04 10:19:38 crc kubenswrapper[4943]: I1204 10:19:38.031179 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5cpfp" podStartSLOduration=2.049872266 podStartE2EDuration="1m7.031159738s" podCreationTimestamp="2025-12-04 10:18:31 +0000 UTC" firstStartedPulling="2025-12-04 10:18:32.502309943 +0000 UTC m=+161.091285811" lastFinishedPulling="2025-12-04 10:19:37.483597405 +0000 UTC m=+226.072573283" observedRunningTime="2025-12-04 10:19:38.027683777 +0000 UTC m=+226.616659645" watchObservedRunningTime="2025-12-04 10:19:38.031159738 +0000 UTC m=+226.620135606" Dec 04 10:19:38 crc kubenswrapper[4943]: I1204 10:19:38.053560 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l7p8s" podStartSLOduration=4.211656094 podStartE2EDuration="1m10.053543524s" podCreationTimestamp="2025-12-04 10:18:28 +0000 UTC" firstStartedPulling="2025-12-04 10:18:31.486406399 +0000 UTC m=+160.075382267" lastFinishedPulling="2025-12-04 10:19:37.328293829 +0000 UTC m=+225.917269697" observedRunningTime="2025-12-04 10:19:38.053491241 +0000 UTC m=+226.642467109" watchObservedRunningTime="2025-12-04 10:19:38.053543524 +0000 UTC m=+226.642519392" Dec 04 10:19:38 crc kubenswrapper[4943]: I1204 10:19:38.107278 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v2fts" podStartSLOduration=2.398583573 podStartE2EDuration="1m6.107254247s" podCreationTimestamp="2025-12-04 10:18:32 +0000 UTC" firstStartedPulling="2025-12-04 10:18:33.561436347 +0000 UTC m=+162.150412225" lastFinishedPulling="2025-12-04 10:19:37.270107031 +0000 UTC m=+225.859082899" observedRunningTime="2025-12-04 10:19:38.104192299 +0000 UTC m=+226.693168177" watchObservedRunningTime="2025-12-04 10:19:38.107254247 +0000 UTC m=+226.696230115" Dec 04 10:19:38 crc kubenswrapper[4943]: I1204 10:19:38.935398 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkx78" event={"ID":"7462ea89-12b3-4d4a-9757-b2e97b190a53","Type":"ContainerStarted","Data":"c317c6884d364eed0a042107e0c08699e71b69e757d5642dfba3a9fa0b882da8"} Dec 04 10:19:38 crc kubenswrapper[4943]: I1204 10:19:38.956034 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lkx78" podStartSLOduration=3.693042652 podStartE2EDuration="1m9.956016579s" podCreationTimestamp="2025-12-04 10:18:29 +0000 UTC" firstStartedPulling="2025-12-04 10:18:31.4709491 +0000 UTC m=+160.059924968" lastFinishedPulling="2025-12-04 10:19:37.733923027 +0000 UTC m=+226.322898895" observedRunningTime="2025-12-04 10:19:38.951566811 +0000 UTC m=+227.540542689" watchObservedRunningTime="2025-12-04 10:19:38.956016579 +0000 UTC m=+227.544992447" Dec 04 10:19:38 crc kubenswrapper[4943]: I1204 10:19:38.993040 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:19:38 crc kubenswrapper[4943]: I1204 10:19:38.993123 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:19:39 crc kubenswrapper[4943]: I1204 10:19:39.048526 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:19:39 crc kubenswrapper[4943]: I1204 10:19:39.235221 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:19:39 crc kubenswrapper[4943]: I1204 10:19:39.235586 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:19:39 crc kubenswrapper[4943]: I1204 10:19:39.472102 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:19:39 crc kubenswrapper[4943]: I1204 10:19:39.472163 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:19:39 crc kubenswrapper[4943]: I1204 10:19:39.762321 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:19:39 crc kubenswrapper[4943]: I1204 10:19:39.985578 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:19:40 crc kubenswrapper[4943]: I1204 10:19:40.282185 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-l7p8s" podUID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" containerName="registry-server" probeResult="failure" output=< Dec 04 10:19:40 crc kubenswrapper[4943]: timeout: failed to connect service ":50051" within 1s Dec 04 10:19:40 crc kubenswrapper[4943]: > Dec 04 10:19:40 crc kubenswrapper[4943]: I1204 10:19:40.521587 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-lkx78" podUID="7462ea89-12b3-4d4a-9757-b2e97b190a53" containerName="registry-server" probeResult="failure" output=< Dec 04 10:19:40 crc kubenswrapper[4943]: timeout: failed to connect service ":50051" within 1s Dec 04 10:19:40 crc kubenswrapper[4943]: > Dec 04 10:19:40 crc kubenswrapper[4943]: I1204 10:19:40.952112 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtr9n" event={"ID":"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa","Type":"ContainerStarted","Data":"b3a7f0acb34017d8a70596513a6daee081b574ee93b80811df18c457c996ae70"} Dec 04 10:19:40 crc kubenswrapper[4943]: I1204 10:19:40.975294 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gtr9n" podStartSLOduration=3.5716151529999998 podStartE2EDuration="1m9.975270075s" podCreationTimestamp="2025-12-04 10:18:31 +0000 UTC" firstStartedPulling="2025-12-04 10:18:33.565441286 +0000 UTC m=+162.154417154" lastFinishedPulling="2025-12-04 10:19:39.969096208 +0000 UTC m=+228.558072076" observedRunningTime="2025-12-04 10:19:40.972518463 +0000 UTC m=+229.561494341" watchObservedRunningTime="2025-12-04 10:19:40.975270075 +0000 UTC m=+229.564245953" Dec 04 10:19:41 crc kubenswrapper[4943]: I1204 10:19:41.570450 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:19:41 crc kubenswrapper[4943]: I1204 10:19:41.570816 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:19:41 crc kubenswrapper[4943]: I1204 10:19:41.632470 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:19:41 crc kubenswrapper[4943]: I1204 10:19:41.803329 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b8cs7"] Dec 04 10:19:41 crc kubenswrapper[4943]: I1204 10:19:41.803663 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b8cs7" podUID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" containerName="registry-server" containerID="cri-o://e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7" gracePeriod=2 Dec 04 10:19:41 crc kubenswrapper[4943]: I1204 10:19:41.963258 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9rd8" event={"ID":"2c9ccdf7-acc0-481a-b1de-5172e16ef22c","Type":"ContainerStarted","Data":"917cf408413d788ff9a009ad32f4722959d1d70a5ce5478a79910ce5ea47277d"} Dec 04 10:19:41 crc kubenswrapper[4943]: I1204 10:19:41.985051 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q9rd8" podStartSLOduration=3.95069787 podStartE2EDuration="1m11.985029468s" podCreationTimestamp="2025-12-04 10:18:30 +0000 UTC" firstStartedPulling="2025-12-04 10:18:32.497745462 +0000 UTC m=+161.086721330" lastFinishedPulling="2025-12-04 10:19:40.53207706 +0000 UTC m=+229.121052928" observedRunningTime="2025-12-04 10:19:41.983551879 +0000 UTC m=+230.572527757" watchObservedRunningTime="2025-12-04 10:19:41.985029468 +0000 UTC m=+230.574005336" Dec 04 10:19:42 crc kubenswrapper[4943]: I1204 10:19:42.266051 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:19:42 crc kubenswrapper[4943]: I1204 10:19:42.266602 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:19:42 crc kubenswrapper[4943]: I1204 10:19:42.475631 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:19:42 crc kubenswrapper[4943]: I1204 10:19:42.475681 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.338398 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gtr9n" podUID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" containerName="registry-server" probeResult="failure" output=< Dec 04 10:19:43 crc kubenswrapper[4943]: timeout: failed to connect service ":50051" within 1s Dec 04 10:19:43 crc kubenswrapper[4943]: > Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.518772 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v2fts" podUID="88878a37-53a8-4edf-8156-ccb1c625ea0d" containerName="registry-server" probeResult="failure" output=< Dec 04 10:19:43 crc kubenswrapper[4943]: timeout: failed to connect service ":50051" within 1s Dec 04 10:19:43 crc kubenswrapper[4943]: > Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.547664 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.748900 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-catalog-content\") pod \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\" (UID: \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\") " Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.749118 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-utilities\") pod \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\" (UID: \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\") " Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.749291 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wzkd\" (UniqueName: \"kubernetes.io/projected/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-kube-api-access-8wzkd\") pod \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\" (UID: \"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565\") " Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.749847 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-utilities" (OuterVolumeSpecName: "utilities") pod "8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" (UID: "8c03ef2d-4f6e-48c5-b6f9-825fc02b8565"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.754399 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-kube-api-access-8wzkd" (OuterVolumeSpecName: "kube-api-access-8wzkd") pod "8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" (UID: "8c03ef2d-4f6e-48c5-b6f9-825fc02b8565"). InnerVolumeSpecName "kube-api-access-8wzkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.804745 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" (UID: "8c03ef2d-4f6e-48c5-b6f9-825fc02b8565"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.850770 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.850811 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.850825 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wzkd\" (UniqueName: \"kubernetes.io/projected/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565-kube-api-access-8wzkd\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.974278 4943 generic.go:334] "Generic (PLEG): container finished" podID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" containerID="e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7" exitCode=0 Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.974328 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8cs7" event={"ID":"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565","Type":"ContainerDied","Data":"e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7"} Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.974371 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8cs7" event={"ID":"8c03ef2d-4f6e-48c5-b6f9-825fc02b8565","Type":"ContainerDied","Data":"837a4a83d35517c20078f0044d61279b17c3c40925e9bd000baaa9b7fc49e565"} Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.974388 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b8cs7" Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.974458 4943 scope.go:117] "RemoveContainer" containerID="e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7" Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.994113 4943 scope.go:117] "RemoveContainer" containerID="c969ab58fdcc2f9af8442794c530b6f6c0394ced30f42cad776b71daf4e135ce" Dec 04 10:19:43 crc kubenswrapper[4943]: I1204 10:19:43.998983 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b8cs7"] Dec 04 10:19:44 crc kubenswrapper[4943]: I1204 10:19:44.005878 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b8cs7"] Dec 04 10:19:44 crc kubenswrapper[4943]: I1204 10:19:44.038388 4943 scope.go:117] "RemoveContainer" containerID="7b3a256df2bb19c6f2d22e11c0d4a1f1e5f4a4b0e53e2588c1ec6c403b9bbc6c" Dec 04 10:19:44 crc kubenswrapper[4943]: I1204 10:19:44.052773 4943 scope.go:117] "RemoveContainer" containerID="e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7" Dec 04 10:19:44 crc kubenswrapper[4943]: E1204 10:19:44.057771 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7\": container with ID starting with e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7 not found: ID does not exist" containerID="e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7" Dec 04 10:19:44 crc kubenswrapper[4943]: I1204 10:19:44.057845 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7"} err="failed to get container status \"e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7\": rpc error: code = NotFound desc = could not find container \"e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7\": container with ID starting with e0ba603eb489aece5ae561b3e39c77a1f084f0e017b470d382a613106749a3b7 not found: ID does not exist" Dec 04 10:19:44 crc kubenswrapper[4943]: I1204 10:19:44.057876 4943 scope.go:117] "RemoveContainer" containerID="c969ab58fdcc2f9af8442794c530b6f6c0394ced30f42cad776b71daf4e135ce" Dec 04 10:19:44 crc kubenswrapper[4943]: E1204 10:19:44.058320 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c969ab58fdcc2f9af8442794c530b6f6c0394ced30f42cad776b71daf4e135ce\": container with ID starting with c969ab58fdcc2f9af8442794c530b6f6c0394ced30f42cad776b71daf4e135ce not found: ID does not exist" containerID="c969ab58fdcc2f9af8442794c530b6f6c0394ced30f42cad776b71daf4e135ce" Dec 04 10:19:44 crc kubenswrapper[4943]: I1204 10:19:44.058360 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c969ab58fdcc2f9af8442794c530b6f6c0394ced30f42cad776b71daf4e135ce"} err="failed to get container status \"c969ab58fdcc2f9af8442794c530b6f6c0394ced30f42cad776b71daf4e135ce\": rpc error: code = NotFound desc = could not find container \"c969ab58fdcc2f9af8442794c530b6f6c0394ced30f42cad776b71daf4e135ce\": container with ID starting with c969ab58fdcc2f9af8442794c530b6f6c0394ced30f42cad776b71daf4e135ce not found: ID does not exist" Dec 04 10:19:44 crc kubenswrapper[4943]: I1204 10:19:44.058380 4943 scope.go:117] "RemoveContainer" containerID="7b3a256df2bb19c6f2d22e11c0d4a1f1e5f4a4b0e53e2588c1ec6c403b9bbc6c" Dec 04 10:19:44 crc kubenswrapper[4943]: E1204 10:19:44.058673 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b3a256df2bb19c6f2d22e11c0d4a1f1e5f4a4b0e53e2588c1ec6c403b9bbc6c\": container with ID starting with 7b3a256df2bb19c6f2d22e11c0d4a1f1e5f4a4b0e53e2588c1ec6c403b9bbc6c not found: ID does not exist" containerID="7b3a256df2bb19c6f2d22e11c0d4a1f1e5f4a4b0e53e2588c1ec6c403b9bbc6c" Dec 04 10:19:44 crc kubenswrapper[4943]: I1204 10:19:44.058717 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b3a256df2bb19c6f2d22e11c0d4a1f1e5f4a4b0e53e2588c1ec6c403b9bbc6c"} err="failed to get container status \"7b3a256df2bb19c6f2d22e11c0d4a1f1e5f4a4b0e53e2588c1ec6c403b9bbc6c\": rpc error: code = NotFound desc = could not find container \"7b3a256df2bb19c6f2d22e11c0d4a1f1e5f4a4b0e53e2588c1ec6c403b9bbc6c\": container with ID starting with 7b3a256df2bb19c6f2d22e11c0d4a1f1e5f4a4b0e53e2588c1ec6c403b9bbc6c not found: ID does not exist" Dec 04 10:19:44 crc kubenswrapper[4943]: I1204 10:19:44.572032 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" path="/var/lib/kubelet/pods/8c03ef2d-4f6e-48c5-b6f9-825fc02b8565/volumes" Dec 04 10:19:49 crc kubenswrapper[4943]: I1204 10:19:49.318502 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:19:49 crc kubenswrapper[4943]: I1204 10:19:49.362545 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:19:49 crc kubenswrapper[4943]: I1204 10:19:49.516429 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:19:49 crc kubenswrapper[4943]: I1204 10:19:49.565867 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:19:51 crc kubenswrapper[4943]: I1204 10:19:51.428160 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:19:51 crc kubenswrapper[4943]: I1204 10:19:51.428539 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:19:51 crc kubenswrapper[4943]: I1204 10:19:51.474143 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:19:51 crc kubenswrapper[4943]: I1204 10:19:51.619192 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:19:52 crc kubenswrapper[4943]: I1204 10:19:52.073582 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:19:52 crc kubenswrapper[4943]: I1204 10:19:52.196099 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lkx78"] Dec 04 10:19:52 crc kubenswrapper[4943]: I1204 10:19:52.196511 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lkx78" podUID="7462ea89-12b3-4d4a-9757-b2e97b190a53" containerName="registry-server" containerID="cri-o://c317c6884d364eed0a042107e0c08699e71b69e757d5642dfba3a9fa0b882da8" gracePeriod=2 Dec 04 10:19:52 crc kubenswrapper[4943]: I1204 10:19:52.323042 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:19:52 crc kubenswrapper[4943]: I1204 10:19:52.363311 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:19:52 crc kubenswrapper[4943]: I1204 10:19:52.514519 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:19:52 crc kubenswrapper[4943]: I1204 10:19:52.555047 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:19:54 crc kubenswrapper[4943]: I1204 10:19:54.000911 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5cpfp"] Dec 04 10:19:54 crc kubenswrapper[4943]: I1204 10:19:54.002302 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5cpfp" podUID="882e619d-b5ca-4455-abcf-04ee85a45001" containerName="registry-server" containerID="cri-o://5797fb1ce00dfe2ef33e3ec217f72e6ee9473cf5d28d72028fe50dbdc1e92b50" gracePeriod=2 Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.056023 4943 generic.go:334] "Generic (PLEG): container finished" podID="7462ea89-12b3-4d4a-9757-b2e97b190a53" containerID="c317c6884d364eed0a042107e0c08699e71b69e757d5642dfba3a9fa0b882da8" exitCode=0 Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.056369 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkx78" event={"ID":"7462ea89-12b3-4d4a-9757-b2e97b190a53","Type":"ContainerDied","Data":"c317c6884d364eed0a042107e0c08699e71b69e757d5642dfba3a9fa0b882da8"} Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.059608 4943 generic.go:334] "Generic (PLEG): container finished" podID="882e619d-b5ca-4455-abcf-04ee85a45001" containerID="5797fb1ce00dfe2ef33e3ec217f72e6ee9473cf5d28d72028fe50dbdc1e92b50" exitCode=0 Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.059649 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cpfp" event={"ID":"882e619d-b5ca-4455-abcf-04ee85a45001","Type":"ContainerDied","Data":"5797fb1ce00dfe2ef33e3ec217f72e6ee9473cf5d28d72028fe50dbdc1e92b50"} Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.625535 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.738029 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7462ea89-12b3-4d4a-9757-b2e97b190a53-catalog-content\") pod \"7462ea89-12b3-4d4a-9757-b2e97b190a53\" (UID: \"7462ea89-12b3-4d4a-9757-b2e97b190a53\") " Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.738129 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88vlp\" (UniqueName: \"kubernetes.io/projected/7462ea89-12b3-4d4a-9757-b2e97b190a53-kube-api-access-88vlp\") pod \"7462ea89-12b3-4d4a-9757-b2e97b190a53\" (UID: \"7462ea89-12b3-4d4a-9757-b2e97b190a53\") " Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.738273 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7462ea89-12b3-4d4a-9757-b2e97b190a53-utilities\") pod \"7462ea89-12b3-4d4a-9757-b2e97b190a53\" (UID: \"7462ea89-12b3-4d4a-9757-b2e97b190a53\") " Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.739055 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7462ea89-12b3-4d4a-9757-b2e97b190a53-utilities" (OuterVolumeSpecName: "utilities") pod "7462ea89-12b3-4d4a-9757-b2e97b190a53" (UID: "7462ea89-12b3-4d4a-9757-b2e97b190a53"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.743451 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7462ea89-12b3-4d4a-9757-b2e97b190a53-kube-api-access-88vlp" (OuterVolumeSpecName: "kube-api-access-88vlp") pod "7462ea89-12b3-4d4a-9757-b2e97b190a53" (UID: "7462ea89-12b3-4d4a-9757-b2e97b190a53"). InnerVolumeSpecName "kube-api-access-88vlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.769288 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.790493 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7462ea89-12b3-4d4a-9757-b2e97b190a53-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7462ea89-12b3-4d4a-9757-b2e97b190a53" (UID: "7462ea89-12b3-4d4a-9757-b2e97b190a53"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.839641 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88vlp\" (UniqueName: \"kubernetes.io/projected/7462ea89-12b3-4d4a-9757-b2e97b190a53-kube-api-access-88vlp\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.839686 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7462ea89-12b3-4d4a-9757-b2e97b190a53-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.839696 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7462ea89-12b3-4d4a-9757-b2e97b190a53-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.940821 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/882e619d-b5ca-4455-abcf-04ee85a45001-catalog-content\") pod \"882e619d-b5ca-4455-abcf-04ee85a45001\" (UID: \"882e619d-b5ca-4455-abcf-04ee85a45001\") " Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.940938 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/882e619d-b5ca-4455-abcf-04ee85a45001-utilities\") pod \"882e619d-b5ca-4455-abcf-04ee85a45001\" (UID: \"882e619d-b5ca-4455-abcf-04ee85a45001\") " Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.941017 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9mlx\" (UniqueName: \"kubernetes.io/projected/882e619d-b5ca-4455-abcf-04ee85a45001-kube-api-access-g9mlx\") pod \"882e619d-b5ca-4455-abcf-04ee85a45001\" (UID: \"882e619d-b5ca-4455-abcf-04ee85a45001\") " Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.941936 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/882e619d-b5ca-4455-abcf-04ee85a45001-utilities" (OuterVolumeSpecName: "utilities") pod "882e619d-b5ca-4455-abcf-04ee85a45001" (UID: "882e619d-b5ca-4455-abcf-04ee85a45001"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.943520 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/882e619d-b5ca-4455-abcf-04ee85a45001-kube-api-access-g9mlx" (OuterVolumeSpecName: "kube-api-access-g9mlx") pod "882e619d-b5ca-4455-abcf-04ee85a45001" (UID: "882e619d-b5ca-4455-abcf-04ee85a45001"). InnerVolumeSpecName "kube-api-access-g9mlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:19:55 crc kubenswrapper[4943]: I1204 10:19:55.963724 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/882e619d-b5ca-4455-abcf-04ee85a45001-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "882e619d-b5ca-4455-abcf-04ee85a45001" (UID: "882e619d-b5ca-4455-abcf-04ee85a45001"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.042968 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/882e619d-b5ca-4455-abcf-04ee85a45001-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.042999 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/882e619d-b5ca-4455-abcf-04ee85a45001-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.043009 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9mlx\" (UniqueName: \"kubernetes.io/projected/882e619d-b5ca-4455-abcf-04ee85a45001-kube-api-access-g9mlx\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.066677 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkx78" event={"ID":"7462ea89-12b3-4d4a-9757-b2e97b190a53","Type":"ContainerDied","Data":"e70325153e732945492149759474892083912a69056da658c340e603e31a3937"} Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.066716 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lkx78" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.066734 4943 scope.go:117] "RemoveContainer" containerID="c317c6884d364eed0a042107e0c08699e71b69e757d5642dfba3a9fa0b882da8" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.068727 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5cpfp" event={"ID":"882e619d-b5ca-4455-abcf-04ee85a45001","Type":"ContainerDied","Data":"d0f6d12d6ec29f99e80887b71eea6359ba67c2da2f02178077f12db977e6e6f4"} Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.068749 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5cpfp" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.083628 4943 scope.go:117] "RemoveContainer" containerID="9f1ed6045be67d4f4cee9dd1fcd2ee0bcaf577aa8cf3fafd1a57a8c08ad7abbc" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.103836 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lkx78"] Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.106819 4943 scope.go:117] "RemoveContainer" containerID="b4210dd1392942849874397481182a098959240fcbd65af4618b63b875473fc1" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.111645 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lkx78"] Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.115617 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5cpfp"] Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.118341 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5cpfp"] Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.124186 4943 scope.go:117] "RemoveContainer" containerID="5797fb1ce00dfe2ef33e3ec217f72e6ee9473cf5d28d72028fe50dbdc1e92b50" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.137282 4943 scope.go:117] "RemoveContainer" containerID="4bd7da92ef80a5c2b0228d2ffea4e81eb50c8722b16fb866f7d58fee35ee4241" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.153451 4943 scope.go:117] "RemoveContainer" containerID="ac4926d1d72f057d1be6cb8d8c34755069890619a95a035abdf35e6502fd5adb" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.575458 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7462ea89-12b3-4d4a-9757-b2e97b190a53" path="/var/lib/kubelet/pods/7462ea89-12b3-4d4a-9757-b2e97b190a53/volumes" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.576404 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="882e619d-b5ca-4455-abcf-04ee85a45001" path="/var/lib/kubelet/pods/882e619d-b5ca-4455-abcf-04ee85a45001/volumes" Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.600481 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v2fts"] Dec 04 10:19:56 crc kubenswrapper[4943]: I1204 10:19:56.600932 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v2fts" podUID="88878a37-53a8-4edf-8156-ccb1c625ea0d" containerName="registry-server" containerID="cri-o://d5b69a44d50a9d99366606e39180a6bb372692b9f3f87013f6bc16a7dddbdacb" gracePeriod=2 Dec 04 10:19:57 crc kubenswrapper[4943]: I1204 10:19:57.080106 4943 generic.go:334] "Generic (PLEG): container finished" podID="88878a37-53a8-4edf-8156-ccb1c625ea0d" containerID="d5b69a44d50a9d99366606e39180a6bb372692b9f3f87013f6bc16a7dddbdacb" exitCode=0 Dec 04 10:19:57 crc kubenswrapper[4943]: I1204 10:19:57.080141 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2fts" event={"ID":"88878a37-53a8-4edf-8156-ccb1c625ea0d","Type":"ContainerDied","Data":"d5b69a44d50a9d99366606e39180a6bb372692b9f3f87013f6bc16a7dddbdacb"} Dec 04 10:19:57 crc kubenswrapper[4943]: I1204 10:19:57.472791 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:19:57 crc kubenswrapper[4943]: I1204 10:19:57.666185 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88878a37-53a8-4edf-8156-ccb1c625ea0d-catalog-content\") pod \"88878a37-53a8-4edf-8156-ccb1c625ea0d\" (UID: \"88878a37-53a8-4edf-8156-ccb1c625ea0d\") " Dec 04 10:19:57 crc kubenswrapper[4943]: I1204 10:19:57.666398 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bj4cn\" (UniqueName: \"kubernetes.io/projected/88878a37-53a8-4edf-8156-ccb1c625ea0d-kube-api-access-bj4cn\") pod \"88878a37-53a8-4edf-8156-ccb1c625ea0d\" (UID: \"88878a37-53a8-4edf-8156-ccb1c625ea0d\") " Dec 04 10:19:57 crc kubenswrapper[4943]: I1204 10:19:57.666442 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88878a37-53a8-4edf-8156-ccb1c625ea0d-utilities\") pod \"88878a37-53a8-4edf-8156-ccb1c625ea0d\" (UID: \"88878a37-53a8-4edf-8156-ccb1c625ea0d\") " Dec 04 10:19:57 crc kubenswrapper[4943]: I1204 10:19:57.668244 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88878a37-53a8-4edf-8156-ccb1c625ea0d-utilities" (OuterVolumeSpecName: "utilities") pod "88878a37-53a8-4edf-8156-ccb1c625ea0d" (UID: "88878a37-53a8-4edf-8156-ccb1c625ea0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:19:57 crc kubenswrapper[4943]: I1204 10:19:57.669703 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88878a37-53a8-4edf-8156-ccb1c625ea0d-kube-api-access-bj4cn" (OuterVolumeSpecName: "kube-api-access-bj4cn") pod "88878a37-53a8-4edf-8156-ccb1c625ea0d" (UID: "88878a37-53a8-4edf-8156-ccb1c625ea0d"). InnerVolumeSpecName "kube-api-access-bj4cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:19:57 crc kubenswrapper[4943]: I1204 10:19:57.768066 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bj4cn\" (UniqueName: \"kubernetes.io/projected/88878a37-53a8-4edf-8156-ccb1c625ea0d-kube-api-access-bj4cn\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:57 crc kubenswrapper[4943]: I1204 10:19:57.768106 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88878a37-53a8-4edf-8156-ccb1c625ea0d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:57 crc kubenswrapper[4943]: I1204 10:19:57.784668 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88878a37-53a8-4edf-8156-ccb1c625ea0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88878a37-53a8-4edf-8156-ccb1c625ea0d" (UID: "88878a37-53a8-4edf-8156-ccb1c625ea0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:19:57 crc kubenswrapper[4943]: I1204 10:19:57.869580 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88878a37-53a8-4edf-8156-ccb1c625ea0d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.095547 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v2fts" event={"ID":"88878a37-53a8-4edf-8156-ccb1c625ea0d","Type":"ContainerDied","Data":"0652ffb167a3ede5e7c081177adf12fda44c3f96cbd63eb951faa15b44bd646b"} Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.095621 4943 scope.go:117] "RemoveContainer" containerID="d5b69a44d50a9d99366606e39180a6bb372692b9f3f87013f6bc16a7dddbdacb" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.095744 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v2fts" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.116513 4943 scope.go:117] "RemoveContainer" containerID="0c4b6b95cfea51fac5c1a00f12711dae33c257ea8138389d66d07fa2dd9cdd2f" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.135275 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v2fts"] Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.141015 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v2fts"] Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.146543 4943 scope.go:117] "RemoveContainer" containerID="7f4f9a4c8e80138a48303aada99e793b4ef7f547cac69c30c5aed322b513dd4d" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.571941 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88878a37-53a8-4edf-8156-ccb1c625ea0d" path="/var/lib/kubelet/pods/88878a37-53a8-4edf-8156-ccb1c625ea0d/volumes" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.693322 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wkldv"] Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910113 4943 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910501 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" containerName="extract-utilities" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910538 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" containerName="extract-utilities" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910552 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="882e619d-b5ca-4455-abcf-04ee85a45001" containerName="registry-server" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910560 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="882e619d-b5ca-4455-abcf-04ee85a45001" containerName="registry-server" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910573 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7462ea89-12b3-4d4a-9757-b2e97b190a53" containerName="extract-utilities" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910580 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7462ea89-12b3-4d4a-9757-b2e97b190a53" containerName="extract-utilities" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910594 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88878a37-53a8-4edf-8156-ccb1c625ea0d" containerName="registry-server" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910601 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="88878a37-53a8-4edf-8156-ccb1c625ea0d" containerName="registry-server" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910613 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="882e619d-b5ca-4455-abcf-04ee85a45001" containerName="extract-content" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910620 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="882e619d-b5ca-4455-abcf-04ee85a45001" containerName="extract-content" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910633 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e67b367b-a25a-452b-9e9b-3af2cdaef856" containerName="pruner" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910640 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e67b367b-a25a-452b-9e9b-3af2cdaef856" containerName="pruner" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910650 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" containerName="registry-server" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910657 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" containerName="registry-server" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910666 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="882e619d-b5ca-4455-abcf-04ee85a45001" containerName="extract-utilities" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910674 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="882e619d-b5ca-4455-abcf-04ee85a45001" containerName="extract-utilities" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910687 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7462ea89-12b3-4d4a-9757-b2e97b190a53" containerName="extract-content" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910694 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7462ea89-12b3-4d4a-9757-b2e97b190a53" containerName="extract-content" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910704 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88878a37-53a8-4edf-8156-ccb1c625ea0d" containerName="extract-utilities" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910713 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="88878a37-53a8-4edf-8156-ccb1c625ea0d" containerName="extract-utilities" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910723 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88878a37-53a8-4edf-8156-ccb1c625ea0d" containerName="extract-content" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910731 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="88878a37-53a8-4edf-8156-ccb1c625ea0d" containerName="extract-content" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910739 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7462ea89-12b3-4d4a-9757-b2e97b190a53" containerName="registry-server" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910746 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7462ea89-12b3-4d4a-9757-b2e97b190a53" containerName="registry-server" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.910758 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" containerName="extract-content" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910765 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" containerName="extract-content" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910944 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="882e619d-b5ca-4455-abcf-04ee85a45001" containerName="registry-server" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910964 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="e67b367b-a25a-452b-9e9b-3af2cdaef856" containerName="pruner" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910978 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="88878a37-53a8-4edf-8156-ccb1c625ea0d" containerName="registry-server" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.910994 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c03ef2d-4f6e-48c5-b6f9-825fc02b8565" containerName="registry-server" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.911003 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="7462ea89-12b3-4d4a-9757-b2e97b190a53" containerName="registry-server" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.911543 4943 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.911748 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.912190 4943 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.912374 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.912395 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.912407 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.912415 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.912427 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.912435 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.912444 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.912452 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.912466 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.912475 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 10:19:58 crc kubenswrapper[4943]: E1204 10:19:58.912488 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.912496 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.912618 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.912630 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.912641 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.912655 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.912666 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.914127 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9" gracePeriod=15 Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.914163 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf" gracePeriod=15 Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.914271 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc" gracePeriod=15 Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.914297 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00" gracePeriod=15 Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.914313 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465" gracePeriod=15 Dec 04 10:19:58 crc kubenswrapper[4943]: I1204 10:19:58.942397 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.037995 4943 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.038076 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.087235 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.087322 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.087391 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.087426 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.087474 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.087504 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.087703 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.087763 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.111648 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.113549 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00" exitCode=0 Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.113681 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf" exitCode=0 Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.113782 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc" exitCode=0 Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.113873 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465" exitCode=2 Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189464 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189529 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189601 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189633 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189686 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189726 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189730 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189770 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189810 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189811 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189851 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189671 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189945 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.189975 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.190001 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.190039 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: I1204 10:19:59.239627 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:19:59 crc kubenswrapper[4943]: W1204 10:19:59.278104 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-395579052b69d674ed4ba0b3002b0b92dcb4caa2deac9a9358e84126c2fc304e WatchSource:0}: Error finding container 395579052b69d674ed4ba0b3002b0b92dcb4caa2deac9a9358e84126c2fc304e: Status 404 returned error can't find the container with id 395579052b69d674ed4ba0b3002b0b92dcb4caa2deac9a9358e84126c2fc304e Dec 04 10:19:59 crc kubenswrapper[4943]: E1204 10:19:59.282824 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dfbe72b420168 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 10:19:59.28086564 +0000 UTC m=+247.869841508,LastTimestamp:2025-12-04 10:19:59.28086564 +0000 UTC m=+247.869841508,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 10:20:00 crc kubenswrapper[4943]: I1204 10:20:00.122358 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"56df8d5eaa2b4f0005b3f6f90fdedde359df1b8251f1493700d8bf8fae0ccac6"} Dec 04 10:20:00 crc kubenswrapper[4943]: I1204 10:20:00.122995 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"395579052b69d674ed4ba0b3002b0b92dcb4caa2deac9a9358e84126c2fc304e"} Dec 04 10:20:00 crc kubenswrapper[4943]: I1204 10:20:00.123123 4943 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:00 crc kubenswrapper[4943]: I1204 10:20:00.123538 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:00 crc kubenswrapper[4943]: I1204 10:20:00.125374 4943 generic.go:334] "Generic (PLEG): container finished" podID="1683641e-8c40-4be2-b18b-af072d514b99" containerID="3da7a9db0af1031e9b446179144f5cf2cdde42a164f0c01ab6baf16c7e4bcafb" exitCode=0 Dec 04 10:20:00 crc kubenswrapper[4943]: I1204 10:20:00.125449 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1683641e-8c40-4be2-b18b-af072d514b99","Type":"ContainerDied","Data":"3da7a9db0af1031e9b446179144f5cf2cdde42a164f0c01ab6baf16c7e4bcafb"} Dec 04 10:20:00 crc kubenswrapper[4943]: I1204 10:20:00.125946 4943 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:00 crc kubenswrapper[4943]: I1204 10:20:00.126260 4943 status_manager.go:851] "Failed to get status for pod" podUID="1683641e-8c40-4be2-b18b-af072d514b99" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:00 crc kubenswrapper[4943]: I1204 10:20:00.126547 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:00 crc kubenswrapper[4943]: E1204 10:20:00.272271 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dfbe72b420168 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 10:19:59.28086564 +0000 UTC m=+247.869841508,LastTimestamp:2025-12-04 10:19:59.28086564 +0000 UTC m=+247.869841508,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.291553 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.292453 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.293156 4943 status_manager.go:851] "Failed to get status for pod" podUID="1683641e-8c40-4be2-b18b-af072d514b99" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.293585 4943 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.293880 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.353853 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.354401 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.354752 4943 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.355191 4943 status_manager.go:851] "Failed to get status for pod" podUID="1683641e-8c40-4be2-b18b-af072d514b99" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.418019 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.418099 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.418135 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.418329 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.418318 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.418434 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.418824 4943 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.419054 4943 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.419098 4943 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.520635 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1683641e-8c40-4be2-b18b-af072d514b99-kubelet-dir\") pod \"1683641e-8c40-4be2-b18b-af072d514b99\" (UID: \"1683641e-8c40-4be2-b18b-af072d514b99\") " Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.520745 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1683641e-8c40-4be2-b18b-af072d514b99-kube-api-access\") pod \"1683641e-8c40-4be2-b18b-af072d514b99\" (UID: \"1683641e-8c40-4be2-b18b-af072d514b99\") " Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.520826 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1683641e-8c40-4be2-b18b-af072d514b99-var-lock\") pod \"1683641e-8c40-4be2-b18b-af072d514b99\" (UID: \"1683641e-8c40-4be2-b18b-af072d514b99\") " Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.521019 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1683641e-8c40-4be2-b18b-af072d514b99-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1683641e-8c40-4be2-b18b-af072d514b99" (UID: "1683641e-8c40-4be2-b18b-af072d514b99"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.521099 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1683641e-8c40-4be2-b18b-af072d514b99-var-lock" (OuterVolumeSpecName: "var-lock") pod "1683641e-8c40-4be2-b18b-af072d514b99" (UID: "1683641e-8c40-4be2-b18b-af072d514b99"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.521499 4943 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1683641e-8c40-4be2-b18b-af072d514b99-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.521530 4943 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1683641e-8c40-4be2-b18b-af072d514b99-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.528933 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1683641e-8c40-4be2-b18b-af072d514b99-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1683641e-8c40-4be2-b18b-af072d514b99" (UID: "1683641e-8c40-4be2-b18b-af072d514b99"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:20:01 crc kubenswrapper[4943]: I1204 10:20:01.624354 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1683641e-8c40-4be2-b18b-af072d514b99-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.140944 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"1683641e-8c40-4be2-b18b-af072d514b99","Type":"ContainerDied","Data":"dffaeb658bbcd012870dbfb4e007a9301a7590570dd20e740742a12d0d68bed9"} Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.141036 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dffaeb658bbcd012870dbfb4e007a9301a7590570dd20e740742a12d0d68bed9" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.141137 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.146961 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.147696 4943 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9" exitCode=0 Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.147830 4943 scope.go:117] "RemoveContainer" containerID="f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.148658 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.157658 4943 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.158616 4943 status_manager.go:851] "Failed to get status for pod" podUID="1683641e-8c40-4be2-b18b-af072d514b99" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.159023 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.167004 4943 status_manager.go:851] "Failed to get status for pod" podUID="1683641e-8c40-4be2-b18b-af072d514b99" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.167184 4943 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.167368 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.167924 4943 scope.go:117] "RemoveContainer" containerID="5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.185327 4943 scope.go:117] "RemoveContainer" containerID="fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.201006 4943 scope.go:117] "RemoveContainer" containerID="ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.214585 4943 scope.go:117] "RemoveContainer" containerID="236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.267114 4943 scope.go:117] "RemoveContainer" containerID="66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.289067 4943 scope.go:117] "RemoveContainer" containerID="f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00" Dec 04 10:20:02 crc kubenswrapper[4943]: E1204 10:20:02.289575 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\": container with ID starting with f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00 not found: ID does not exist" containerID="f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.289647 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00"} err="failed to get container status \"f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\": rpc error: code = NotFound desc = could not find container \"f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00\": container with ID starting with f8a84784b83f42eea88c93288a675453dc6a04e4dfab52426a72f07ece7adf00 not found: ID does not exist" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.289724 4943 scope.go:117] "RemoveContainer" containerID="5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf" Dec 04 10:20:02 crc kubenswrapper[4943]: E1204 10:20:02.290130 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\": container with ID starting with 5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf not found: ID does not exist" containerID="5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.290159 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf"} err="failed to get container status \"5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\": rpc error: code = NotFound desc = could not find container \"5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf\": container with ID starting with 5522f364b5cacaaca12c24dedfd23e350021869fcd14532a321926cee6bb67bf not found: ID does not exist" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.290173 4943 scope.go:117] "RemoveContainer" containerID="fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc" Dec 04 10:20:02 crc kubenswrapper[4943]: E1204 10:20:02.291043 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\": container with ID starting with fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc not found: ID does not exist" containerID="fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.291120 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc"} err="failed to get container status \"fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\": rpc error: code = NotFound desc = could not find container \"fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc\": container with ID starting with fe7a7f4ac377a2f32fc07614379e5b81702aca5c56c1f62430c00114c782f2fc not found: ID does not exist" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.291148 4943 scope.go:117] "RemoveContainer" containerID="ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465" Dec 04 10:20:02 crc kubenswrapper[4943]: E1204 10:20:02.292320 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\": container with ID starting with ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465 not found: ID does not exist" containerID="ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.292368 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465"} err="failed to get container status \"ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\": rpc error: code = NotFound desc = could not find container \"ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465\": container with ID starting with ed5cfec5233a62b479687b77cdd6022358be222b6a1e117795147980eb0a2465 not found: ID does not exist" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.292402 4943 scope.go:117] "RemoveContainer" containerID="236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9" Dec 04 10:20:02 crc kubenswrapper[4943]: E1204 10:20:02.293342 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\": container with ID starting with 236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9 not found: ID does not exist" containerID="236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.293367 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9"} err="failed to get container status \"236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\": rpc error: code = NotFound desc = could not find container \"236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9\": container with ID starting with 236011e4317527c9c513ba82bfba305c566e12ef8a3833f9bd2a30eb25e12ee9 not found: ID does not exist" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.293382 4943 scope.go:117] "RemoveContainer" containerID="66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0" Dec 04 10:20:02 crc kubenswrapper[4943]: E1204 10:20:02.293669 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\": container with ID starting with 66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0 not found: ID does not exist" containerID="66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.293719 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0"} err="failed to get container status \"66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\": rpc error: code = NotFound desc = could not find container \"66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0\": container with ID starting with 66d7a6ef6657ee3c77fa977d71890e27803e79e10dea8cf90790356871e656b0 not found: ID does not exist" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.568278 4943 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.568929 4943 status_manager.go:851] "Failed to get status for pod" podUID="1683641e-8c40-4be2-b18b-af072d514b99" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.569176 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:02 crc kubenswrapper[4943]: I1204 10:20:02.573615 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 04 10:20:03 crc kubenswrapper[4943]: E1204 10:20:03.793873 4943 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:03 crc kubenswrapper[4943]: E1204 10:20:03.794364 4943 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:03 crc kubenswrapper[4943]: E1204 10:20:03.794724 4943 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:03 crc kubenswrapper[4943]: E1204 10:20:03.795041 4943 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:03 crc kubenswrapper[4943]: E1204 10:20:03.795418 4943 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:03 crc kubenswrapper[4943]: I1204 10:20:03.795478 4943 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 04 10:20:03 crc kubenswrapper[4943]: E1204 10:20:03.795815 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="200ms" Dec 04 10:20:03 crc kubenswrapper[4943]: E1204 10:20:03.996414 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="400ms" Dec 04 10:20:04 crc kubenswrapper[4943]: E1204 10:20:04.398321 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="800ms" Dec 04 10:20:05 crc kubenswrapper[4943]: E1204 10:20:05.199439 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="1.6s" Dec 04 10:20:06 crc kubenswrapper[4943]: E1204 10:20:06.800966 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="3.2s" Dec 04 10:20:10 crc kubenswrapper[4943]: E1204 10:20:10.002022 4943 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.150:6443: connect: connection refused" interval="6.4s" Dec 04 10:20:10 crc kubenswrapper[4943]: E1204 10:20:10.273960 4943 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.150:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dfbe72b420168 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 10:19:59.28086564 +0000 UTC m=+247.869841508,LastTimestamp:2025-12-04 10:19:59.28086564 +0000 UTC m=+247.869841508,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 10:20:10 crc kubenswrapper[4943]: I1204 10:20:10.564856 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:10 crc kubenswrapper[4943]: I1204 10:20:10.566263 4943 status_manager.go:851] "Failed to get status for pod" podUID="1683641e-8c40-4be2-b18b-af072d514b99" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:10 crc kubenswrapper[4943]: I1204 10:20:10.566877 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:10 crc kubenswrapper[4943]: I1204 10:20:10.581406 4943 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e8c3544b-029d-48af-862b-a4c40406cfc7" Dec 04 10:20:10 crc kubenswrapper[4943]: I1204 10:20:10.581502 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e8c3544b-029d-48af-862b-a4c40406cfc7" Dec 04 10:20:10 crc kubenswrapper[4943]: E1204 10:20:10.582179 4943 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:10 crc kubenswrapper[4943]: I1204 10:20:10.583024 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.213124 4943 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="0c7ad63c03c7f1e13813710a4fa5c5147c31741f7dc1d41283b8a5083293cf3f" exitCode=0 Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.213289 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"0c7ad63c03c7f1e13813710a4fa5c5147c31741f7dc1d41283b8a5083293cf3f"} Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.213558 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3e94f70601078758a35ff80c64fad27f241d1b5847788a69c28aa04c8de19477"} Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.213895 4943 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e8c3544b-029d-48af-862b-a4c40406cfc7" Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.213914 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e8c3544b-029d-48af-862b-a4c40406cfc7" Dec 04 10:20:11 crc kubenswrapper[4943]: E1204 10:20:11.214402 4943 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.214442 4943 status_manager.go:851] "Failed to get status for pod" podUID="1683641e-8c40-4be2-b18b-af072d514b99" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.214841 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.217545 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.217604 4943 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf" exitCode=1 Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.217636 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf"} Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.218114 4943 scope.go:117] "RemoveContainer" containerID="c397988686b29bc79196bd67bd017867735089b050a43a6aac2f81e1059730cf" Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.218472 4943 status_manager.go:851] "Failed to get status for pod" podUID="1683641e-8c40-4be2-b18b-af072d514b99" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.219034 4943 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.219519 4943 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.150:6443: connect: connection refused" Dec 04 10:20:11 crc kubenswrapper[4943]: I1204 10:20:11.516060 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:20:12 crc kubenswrapper[4943]: I1204 10:20:12.226143 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"34292deb737b76d60dda8f15d2ff1bbe46e442d8e39e90df4f90c062a9261d6b"} Dec 04 10:20:12 crc kubenswrapper[4943]: I1204 10:20:12.226450 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"dd7cc5bca207dad61e5151220b436f60153ed7a20583b03054b5891de75b655d"} Dec 04 10:20:12 crc kubenswrapper[4943]: I1204 10:20:12.226460 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"303fc3168f0e56ee9b170a8c24bb281e57189f73c81eb66f15eb00dd4daa969f"} Dec 04 10:20:12 crc kubenswrapper[4943]: I1204 10:20:12.226468 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f24fe60886015be6a7962753d6c12fb5c8de88dfda7c81b35101313bf8251051"} Dec 04 10:20:12 crc kubenswrapper[4943]: I1204 10:20:12.230459 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 10:20:12 crc kubenswrapper[4943]: I1204 10:20:12.230530 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7b806c4f2539d62d4959782880d3b37361e4c411c73096d44f208f34f0b2b8f4"} Dec 04 10:20:13 crc kubenswrapper[4943]: I1204 10:20:13.240456 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ab653e4b0d8ff3428bd427860b62a1ce99170e212151dbcfd1d03d962b3be8ee"} Dec 04 10:20:13 crc kubenswrapper[4943]: I1204 10:20:13.241420 4943 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e8c3544b-029d-48af-862b-a4c40406cfc7" Dec 04 10:20:13 crc kubenswrapper[4943]: I1204 10:20:13.241444 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e8c3544b-029d-48af-862b-a4c40406cfc7" Dec 04 10:20:13 crc kubenswrapper[4943]: I1204 10:20:13.241705 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:13 crc kubenswrapper[4943]: I1204 10:20:13.585852 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:20:15 crc kubenswrapper[4943]: I1204 10:20:15.583349 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:15 crc kubenswrapper[4943]: I1204 10:20:15.583673 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:15 crc kubenswrapper[4943]: I1204 10:20:15.596703 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:18 crc kubenswrapper[4943]: I1204 10:20:18.250562 4943 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:18 crc kubenswrapper[4943]: I1204 10:20:18.267336 4943 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e8c3544b-029d-48af-862b-a4c40406cfc7" Dec 04 10:20:18 crc kubenswrapper[4943]: I1204 10:20:18.267373 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e8c3544b-029d-48af-862b-a4c40406cfc7" Dec 04 10:20:18 crc kubenswrapper[4943]: I1204 10:20:18.277236 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:18 crc kubenswrapper[4943]: I1204 10:20:18.283475 4943 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d9b35a18-2183-4c5b-9495-2ac230ba9d01" Dec 04 10:20:19 crc kubenswrapper[4943]: I1204 10:20:19.271732 4943 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e8c3544b-029d-48af-862b-a4c40406cfc7" Dec 04 10:20:19 crc kubenswrapper[4943]: I1204 10:20:19.271765 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e8c3544b-029d-48af-862b-a4c40406cfc7" Dec 04 10:20:21 crc kubenswrapper[4943]: I1204 10:20:21.516316 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:20:21 crc kubenswrapper[4943]: I1204 10:20:21.520855 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:20:22 crc kubenswrapper[4943]: I1204 10:20:22.297594 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 10:20:22 crc kubenswrapper[4943]: I1204 10:20:22.580725 4943 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="d9b35a18-2183-4c5b-9495-2ac230ba9d01" Dec 04 10:20:23 crc kubenswrapper[4943]: I1204 10:20:23.724868 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" podUID="14bd93bd-8f38-48c3-9508-6c1ab9e65628" containerName="oauth-openshift" containerID="cri-o://aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1" gracePeriod=15 Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.119347 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.251685 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-serving-cert\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.251749 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2s7d\" (UniqueName: \"kubernetes.io/projected/14bd93bd-8f38-48c3-9508-6c1ab9e65628-kube-api-access-g2s7d\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.251788 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14bd93bd-8f38-48c3-9508-6c1ab9e65628-audit-dir\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.251816 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-router-certs\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.251845 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-ocp-branding-template\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.251885 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-error\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.251919 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-session\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.251943 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14bd93bd-8f38-48c3-9508-6c1ab9e65628-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.251969 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-trusted-ca-bundle\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.252043 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-idp-0-file-data\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.252079 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-audit-policies\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.252098 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-provider-selection\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.252114 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-login\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.252710 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.252729 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.253188 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-service-ca\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.253246 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-cliconfig\") pod \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\" (UID: \"14bd93bd-8f38-48c3-9508-6c1ab9e65628\") " Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.253497 4943 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.253515 4943 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/14bd93bd-8f38-48c3-9508-6c1ab9e65628-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.253526 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.253949 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.253983 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.270393 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.270768 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.271083 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.271405 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.271594 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.271886 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.271953 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.272212 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.272589 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14bd93bd-8f38-48c3-9508-6c1ab9e65628-kube-api-access-g2s7d" (OuterVolumeSpecName: "kube-api-access-g2s7d") pod "14bd93bd-8f38-48c3-9508-6c1ab9e65628" (UID: "14bd93bd-8f38-48c3-9508-6c1ab9e65628"). InnerVolumeSpecName "kube-api-access-g2s7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.303247 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" event={"ID":"14bd93bd-8f38-48c3-9508-6c1ab9e65628","Type":"ContainerDied","Data":"aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1"} Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.303330 4943 scope.go:117] "RemoveContainer" containerID="aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.303331 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.303184 4943 generic.go:334] "Generic (PLEG): container finished" podID="14bd93bd-8f38-48c3-9508-6c1ab9e65628" containerID="aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1" exitCode=0 Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.303494 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wkldv" event={"ID":"14bd93bd-8f38-48c3-9508-6c1ab9e65628","Type":"ContainerDied","Data":"3a4b6bfeeae45122b65be01796496703f4ab683dbd5090b41401134d15c2ee35"} Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.328190 4943 scope.go:117] "RemoveContainer" containerID="aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1" Dec 04 10:20:24 crc kubenswrapper[4943]: E1204 10:20:24.329537 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1\": container with ID starting with aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1 not found: ID does not exist" containerID="aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.329601 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1"} err="failed to get container status \"aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1\": rpc error: code = NotFound desc = could not find container \"aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1\": container with ID starting with aea6755441998861ba9e40379b3a6b97c649201197a76890f491286345e7c3d1 not found: ID does not exist" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.354253 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.354288 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.354303 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.354312 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.354323 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.354331 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.354387 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2s7d\" (UniqueName: \"kubernetes.io/projected/14bd93bd-8f38-48c3-9508-6c1ab9e65628-kube-api-access-g2s7d\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.354402 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.354417 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.354429 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:24 crc kubenswrapper[4943]: I1204 10:20:24.354440 4943 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/14bd93bd-8f38-48c3-9508-6c1ab9e65628-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:28 crc kubenswrapper[4943]: I1204 10:20:28.283764 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 10:20:29 crc kubenswrapper[4943]: I1204 10:20:29.534272 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 10:20:29 crc kubenswrapper[4943]: I1204 10:20:29.739617 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 10:20:29 crc kubenswrapper[4943]: I1204 10:20:29.943142 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 10:20:30 crc kubenswrapper[4943]: I1204 10:20:30.272502 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 10:20:30 crc kubenswrapper[4943]: I1204 10:20:30.462455 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 10:20:30 crc kubenswrapper[4943]: I1204 10:20:30.954499 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 10:20:31 crc kubenswrapper[4943]: I1204 10:20:31.064573 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 10:20:31 crc kubenswrapper[4943]: I1204 10:20:31.135120 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 10:20:31 crc kubenswrapper[4943]: I1204 10:20:31.291277 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 10:20:31 crc kubenswrapper[4943]: I1204 10:20:31.314367 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 10:20:31 crc kubenswrapper[4943]: I1204 10:20:31.370640 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 10:20:31 crc kubenswrapper[4943]: I1204 10:20:31.389821 4943 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 10:20:31 crc kubenswrapper[4943]: I1204 10:20:31.459759 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 10:20:31 crc kubenswrapper[4943]: I1204 10:20:31.635833 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 10:20:31 crc kubenswrapper[4943]: I1204 10:20:31.806342 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 10:20:31 crc kubenswrapper[4943]: I1204 10:20:31.955184 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 10:20:31 crc kubenswrapper[4943]: I1204 10:20:31.964904 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 10:20:32 crc kubenswrapper[4943]: I1204 10:20:32.022760 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 10:20:32 crc kubenswrapper[4943]: I1204 10:20:32.118162 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 10:20:32 crc kubenswrapper[4943]: I1204 10:20:32.240479 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 10:20:32 crc kubenswrapper[4943]: I1204 10:20:32.479142 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 10:20:32 crc kubenswrapper[4943]: I1204 10:20:32.588826 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 10:20:32 crc kubenswrapper[4943]: I1204 10:20:32.604375 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 10:20:32 crc kubenswrapper[4943]: I1204 10:20:32.805611 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.132991 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.135042 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.158698 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.229746 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.354775 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.397720 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.565153 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.571086 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.620984 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.755608 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.779902 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.807465 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 10:20:33 crc kubenswrapper[4943]: I1204 10:20:33.939694 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.008053 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.061249 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.130169 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.153407 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.293745 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.346869 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.351621 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.407262 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.414063 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.414131 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.496554 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.545045 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.587509 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.671966 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.785520 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 10:20:34 crc kubenswrapper[4943]: I1204 10:20:34.786145 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.043960 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.182871 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.198226 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.256441 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.275103 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.335111 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.381164 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.421631 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.435673 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.467566 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.555364 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.568445 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.624053 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.714551 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.733016 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.887433 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 10:20:35 crc kubenswrapper[4943]: I1204 10:20:35.955133 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.013227 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.025630 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.025632 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.080659 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.103820 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.138925 4943 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.184431 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.206328 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.214047 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.325671 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.350772 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.438626 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.447422 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.461455 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.461896 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.476031 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.557504 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.592415 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.606978 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.660532 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.700346 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.747642 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.868835 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 10:20:36 crc kubenswrapper[4943]: I1204 10:20:36.896027 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.127431 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.347791 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.369108 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.374620 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.385713 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.482380 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.486459 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.579097 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.595105 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.673865 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.716470 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.891310 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.893525 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.906811 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.921481 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.939710 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.957346 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 10:20:37 crc kubenswrapper[4943]: I1204 10:20:37.974629 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.021088 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.042158 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.060627 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.066505 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.121745 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.175540 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.201730 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.208563 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.296370 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.324519 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.406680 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.414318 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.435436 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.453388 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.621703 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.715073 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.759285 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.798379 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.800027 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.835861 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.836510 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.839327 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.881934 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.905517 4943 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.911054 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=40.911020234 podStartE2EDuration="40.911020234s" podCreationTimestamp="2025-12-04 10:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:20:17.910951607 +0000 UTC m=+266.499927475" watchObservedRunningTime="2025-12-04 10:20:38.911020234 +0000 UTC m=+287.499996102" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.912364 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wkldv","openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.912478 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd"] Dec 04 10:20:38 crc kubenswrapper[4943]: E1204 10:20:38.912761 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1683641e-8c40-4be2-b18b-af072d514b99" containerName="installer" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.912792 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1683641e-8c40-4be2-b18b-af072d514b99" containerName="installer" Dec 04 10:20:38 crc kubenswrapper[4943]: E1204 10:20:38.912806 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14bd93bd-8f38-48c3-9508-6c1ab9e65628" containerName="oauth-openshift" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.912812 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="14bd93bd-8f38-48c3-9508-6c1ab9e65628" containerName="oauth-openshift" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.912953 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1683641e-8c40-4be2-b18b-af072d514b99" containerName="installer" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.912973 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="14bd93bd-8f38-48c3-9508-6c1ab9e65628" containerName="oauth-openshift" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.913169 4943 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e8c3544b-029d-48af-862b-a4c40406cfc7" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.913252 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e8c3544b-029d-48af-862b-a4c40406cfc7" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.913675 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.918825 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.921586 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.924027 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.924270 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.924538 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.924903 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.925182 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.925544 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.926400 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.926731 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.927717 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.927805 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.930828 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.935104 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.935586 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.940726 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.944722 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.944696533 podStartE2EDuration="20.944696533s" podCreationTimestamp="2025-12-04 10:20:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:20:38.943026175 +0000 UTC m=+287.532002043" watchObservedRunningTime="2025-12-04 10:20:38.944696533 +0000 UTC m=+287.533672401" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.986678 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.987719 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-session\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.987837 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcrw4\" (UniqueName: \"kubernetes.io/projected/0ac24e88-edb0-4881-b73b-81527279bed9-kube-api-access-qcrw4\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.987866 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-user-template-error\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.987894 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.987913 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-user-template-login\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.987958 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.987979 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.987999 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.988023 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.988099 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0ac24e88-edb0-4881-b73b-81527279bed9-audit-policies\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.988130 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0ac24e88-edb0-4881-b73b-81527279bed9-audit-dir\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.988154 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.988172 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:38 crc kubenswrapper[4943]: I1204 10:20:38.988235 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088560 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-user-template-login\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088614 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088637 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088682 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088705 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088740 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0ac24e88-edb0-4881-b73b-81527279bed9-audit-policies\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088764 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0ac24e88-edb0-4881-b73b-81527279bed9-audit-dir\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088790 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088808 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088830 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088861 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-session\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088877 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcrw4\" (UniqueName: \"kubernetes.io/projected/0ac24e88-edb0-4881-b73b-81527279bed9-kube-api-access-qcrw4\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088906 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-user-template-error\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.088924 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.090124 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0ac24e88-edb0-4881-b73b-81527279bed9-audit-dir\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.091868 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.092188 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.092377 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.092968 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0ac24e88-edb0-4881-b73b-81527279bed9-audit-policies\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.096217 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.096227 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-user-template-error\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.096701 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.097002 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-user-template-login\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.097095 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.097389 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.097487 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-session\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.101093 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0ac24e88-edb0-4881-b73b-81527279bed9-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.115556 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcrw4\" (UniqueName: \"kubernetes.io/projected/0ac24e88-edb0-4881-b73b-81527279bed9-kube-api-access-qcrw4\") pod \"oauth-openshift-5d4f55d7c5-8fjxd\" (UID: \"0ac24e88-edb0-4881-b73b-81527279bed9\") " pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.127577 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.143006 4943 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.178737 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.239759 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.291346 4943 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.370772 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.453062 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.453414 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.468417 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.511895 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.521559 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd"] Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.652854 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.675896 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.683934 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.698428 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.818917 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.843939 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.890269 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.942665 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 10:20:39 crc kubenswrapper[4943]: I1204 10:20:39.967865 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.095390 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.141295 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.159876 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.207715 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.245521 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.270498 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.284539 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.287599 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.397905 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" event={"ID":"0ac24e88-edb0-4881-b73b-81527279bed9","Type":"ContainerStarted","Data":"38e8e79f450a602ad619c5c8d7323ab36299d4d41e5b2db714d04422eb2c56ae"} Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.397957 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" event={"ID":"0ac24e88-edb0-4881-b73b-81527279bed9","Type":"ContainerStarted","Data":"1ffd67000e14c3d20627cb93c14d74ab56f5462fe57ae067ac55b321d242fb77"} Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.398546 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.405549 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.421451 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5d4f55d7c5-8fjxd" podStartSLOduration=42.421427027 podStartE2EDuration="42.421427027s" podCreationTimestamp="2025-12-04 10:19:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:20:40.41942931 +0000 UTC m=+289.008405198" watchObservedRunningTime="2025-12-04 10:20:40.421427027 +0000 UTC m=+289.010402895" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.437944 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.470127 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.573024 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14bd93bd-8f38-48c3-9508-6c1ab9e65628" path="/var/lib/kubelet/pods/14bd93bd-8f38-48c3-9508-6c1ab9e65628/volumes" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.593011 4943 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.593382 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://56df8d5eaa2b4f0005b3f6f90fdedde359df1b8251f1493700d8bf8fae0ccac6" gracePeriod=5 Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.643244 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.690611 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.739893 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.740903 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 10:20:40 crc kubenswrapper[4943]: I1204 10:20:40.977746 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.160904 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.194578 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.232860 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.394953 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.462620 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.533308 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.538210 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.620707 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.655233 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.670179 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.708465 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.882481 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 10:20:41 crc kubenswrapper[4943]: I1204 10:20:41.891287 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.039757 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.064576 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.098778 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.235329 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.250792 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.370622 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.398124 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.596596 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.728346 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.783445 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.835008 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.916005 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.943631 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 10:20:42 crc kubenswrapper[4943]: I1204 10:20:42.978972 4943 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.068720 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.090275 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.123094 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.168368 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.207893 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.212641 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.242550 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.472045 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.526855 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.669394 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.753018 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.761786 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.825774 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.834786 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 10:20:43 crc kubenswrapper[4943]: I1204 10:20:43.909963 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 10:20:44 crc kubenswrapper[4943]: I1204 10:20:44.103401 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 10:20:44 crc kubenswrapper[4943]: I1204 10:20:44.166894 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 10:20:44 crc kubenswrapper[4943]: I1204 10:20:44.180089 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 10:20:44 crc kubenswrapper[4943]: I1204 10:20:44.215799 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 10:20:44 crc kubenswrapper[4943]: I1204 10:20:44.359048 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 10:20:44 crc kubenswrapper[4943]: I1204 10:20:44.490888 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 10:20:44 crc kubenswrapper[4943]: I1204 10:20:44.545763 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 10:20:44 crc kubenswrapper[4943]: I1204 10:20:44.591748 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 10:20:44 crc kubenswrapper[4943]: I1204 10:20:44.851115 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 10:20:44 crc kubenswrapper[4943]: I1204 10:20:44.898495 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 10:20:44 crc kubenswrapper[4943]: I1204 10:20:44.972574 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 10:20:45 crc kubenswrapper[4943]: I1204 10:20:45.194147 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 10:20:45 crc kubenswrapper[4943]: I1204 10:20:45.362474 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 10:20:45 crc kubenswrapper[4943]: I1204 10:20:45.378147 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 10:20:45 crc kubenswrapper[4943]: I1204 10:20:45.868848 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.134243 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.152674 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.176168 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.176263 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.283489 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.375984 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.376042 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.376090 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.376064 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.376142 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.376228 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.376227 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.376240 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.376273 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.376406 4943 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.376421 4943 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.376432 4943 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.376442 4943 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.383499 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.438320 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.438380 4943 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="56df8d5eaa2b4f0005b3f6f90fdedde359df1b8251f1493700d8bf8fae0ccac6" exitCode=137 Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.438446 4943 scope.go:117] "RemoveContainer" containerID="56df8d5eaa2b4f0005b3f6f90fdedde359df1b8251f1493700d8bf8fae0ccac6" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.438478 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.467058 4943 scope.go:117] "RemoveContainer" containerID="56df8d5eaa2b4f0005b3f6f90fdedde359df1b8251f1493700d8bf8fae0ccac6" Dec 04 10:20:46 crc kubenswrapper[4943]: E1204 10:20:46.467686 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56df8d5eaa2b4f0005b3f6f90fdedde359df1b8251f1493700d8bf8fae0ccac6\": container with ID starting with 56df8d5eaa2b4f0005b3f6f90fdedde359df1b8251f1493700d8bf8fae0ccac6 not found: ID does not exist" containerID="56df8d5eaa2b4f0005b3f6f90fdedde359df1b8251f1493700d8bf8fae0ccac6" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.467732 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56df8d5eaa2b4f0005b3f6f90fdedde359df1b8251f1493700d8bf8fae0ccac6"} err="failed to get container status \"56df8d5eaa2b4f0005b3f6f90fdedde359df1b8251f1493700d8bf8fae0ccac6\": rpc error: code = NotFound desc = could not find container \"56df8d5eaa2b4f0005b3f6f90fdedde359df1b8251f1493700d8bf8fae0ccac6\": container with ID starting with 56df8d5eaa2b4f0005b3f6f90fdedde359df1b8251f1493700d8bf8fae0ccac6 not found: ID does not exist" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.477360 4943 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.573930 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.574182 4943 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.585104 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.585140 4943 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="b3c002c2-66c2-4253-978c-6158c154bf42" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.588597 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.588645 4943 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="b3c002c2-66c2-4253-978c-6158c154bf42" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.780181 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 10:20:46 crc kubenswrapper[4943]: I1204 10:20:46.831286 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 10:20:47 crc kubenswrapper[4943]: I1204 10:20:47.369693 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 10:20:47 crc kubenswrapper[4943]: I1204 10:20:47.471790 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 10:20:47 crc kubenswrapper[4943]: I1204 10:20:47.908949 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 10:20:47 crc kubenswrapper[4943]: I1204 10:20:47.985486 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 10:21:08 crc kubenswrapper[4943]: I1204 10:21:08.571538 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2g96c"] Dec 04 10:21:08 crc kubenswrapper[4943]: I1204 10:21:08.572466 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" podUID="8490c962-38c0-4be5-b940-d4c4d08c8a55" containerName="controller-manager" containerID="cri-o://6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a" gracePeriod=30 Dec 04 10:21:08 crc kubenswrapper[4943]: I1204 10:21:08.664751 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9"] Dec 04 10:21:08 crc kubenswrapper[4943]: I1204 10:21:08.664967 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" podUID="58b0ae05-55dc-4238-9163-c8e06974ffca" containerName="route-controller-manager" containerID="cri-o://df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36" gracePeriod=30 Dec 04 10:21:08 crc kubenswrapper[4943]: I1204 10:21:08.973880 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.056462 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.106828 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j72zg\" (UniqueName: \"kubernetes.io/projected/8490c962-38c0-4be5-b940-d4c4d08c8a55-kube-api-access-j72zg\") pod \"8490c962-38c0-4be5-b940-d4c4d08c8a55\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.106911 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-proxy-ca-bundles\") pod \"8490c962-38c0-4be5-b940-d4c4d08c8a55\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.107368 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-config\") pod \"8490c962-38c0-4be5-b940-d4c4d08c8a55\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.107412 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8490c962-38c0-4be5-b940-d4c4d08c8a55-serving-cert\") pod \"8490c962-38c0-4be5-b940-d4c4d08c8a55\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.107437 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-client-ca\") pod \"8490c962-38c0-4be5-b940-d4c4d08c8a55\" (UID: \"8490c962-38c0-4be5-b940-d4c4d08c8a55\") " Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.107465 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58b0ae05-55dc-4238-9163-c8e06974ffca-serving-cert\") pod \"58b0ae05-55dc-4238-9163-c8e06974ffca\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.108708 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-config" (OuterVolumeSpecName: "config") pod "8490c962-38c0-4be5-b940-d4c4d08c8a55" (UID: "8490c962-38c0-4be5-b940-d4c4d08c8a55"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.108709 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-client-ca" (OuterVolumeSpecName: "client-ca") pod "8490c962-38c0-4be5-b940-d4c4d08c8a55" (UID: "8490c962-38c0-4be5-b940-d4c4d08c8a55"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.108729 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8490c962-38c0-4be5-b940-d4c4d08c8a55" (UID: "8490c962-38c0-4be5-b940-d4c4d08c8a55"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.113319 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8490c962-38c0-4be5-b940-d4c4d08c8a55-kube-api-access-j72zg" (OuterVolumeSpecName: "kube-api-access-j72zg") pod "8490c962-38c0-4be5-b940-d4c4d08c8a55" (UID: "8490c962-38c0-4be5-b940-d4c4d08c8a55"). InnerVolumeSpecName "kube-api-access-j72zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.113705 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8490c962-38c0-4be5-b940-d4c4d08c8a55-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8490c962-38c0-4be5-b940-d4c4d08c8a55" (UID: "8490c962-38c0-4be5-b940-d4c4d08c8a55"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.114054 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58b0ae05-55dc-4238-9163-c8e06974ffca-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "58b0ae05-55dc-4238-9163-c8e06974ffca" (UID: "58b0ae05-55dc-4238-9163-c8e06974ffca"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.208407 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6ssz\" (UniqueName: \"kubernetes.io/projected/58b0ae05-55dc-4238-9163-c8e06974ffca-kube-api-access-w6ssz\") pod \"58b0ae05-55dc-4238-9163-c8e06974ffca\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.208489 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58b0ae05-55dc-4238-9163-c8e06974ffca-client-ca\") pod \"58b0ae05-55dc-4238-9163-c8e06974ffca\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.208532 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58b0ae05-55dc-4238-9163-c8e06974ffca-config\") pod \"58b0ae05-55dc-4238-9163-c8e06974ffca\" (UID: \"58b0ae05-55dc-4238-9163-c8e06974ffca\") " Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.208761 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j72zg\" (UniqueName: \"kubernetes.io/projected/8490c962-38c0-4be5-b940-d4c4d08c8a55-kube-api-access-j72zg\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.208782 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.208798 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.208812 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8490c962-38c0-4be5-b940-d4c4d08c8a55-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.208825 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8490c962-38c0-4be5-b940-d4c4d08c8a55-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.208835 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/58b0ae05-55dc-4238-9163-c8e06974ffca-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.209494 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58b0ae05-55dc-4238-9163-c8e06974ffca-config" (OuterVolumeSpecName: "config") pod "58b0ae05-55dc-4238-9163-c8e06974ffca" (UID: "58b0ae05-55dc-4238-9163-c8e06974ffca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.209617 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58b0ae05-55dc-4238-9163-c8e06974ffca-client-ca" (OuterVolumeSpecName: "client-ca") pod "58b0ae05-55dc-4238-9163-c8e06974ffca" (UID: "58b0ae05-55dc-4238-9163-c8e06974ffca"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.211884 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58b0ae05-55dc-4238-9163-c8e06974ffca-kube-api-access-w6ssz" (OuterVolumeSpecName: "kube-api-access-w6ssz") pod "58b0ae05-55dc-4238-9163-c8e06974ffca" (UID: "58b0ae05-55dc-4238-9163-c8e06974ffca"). InnerVolumeSpecName "kube-api-access-w6ssz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.309543 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58b0ae05-55dc-4238-9163-c8e06974ffca-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.309580 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6ssz\" (UniqueName: \"kubernetes.io/projected/58b0ae05-55dc-4238-9163-c8e06974ffca-kube-api-access-w6ssz\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.309590 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/58b0ae05-55dc-4238-9163-c8e06974ffca-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.566619 4943 generic.go:334] "Generic (PLEG): container finished" podID="58b0ae05-55dc-4238-9163-c8e06974ffca" containerID="df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36" exitCode=0 Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.566739 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" event={"ID":"58b0ae05-55dc-4238-9163-c8e06974ffca","Type":"ContainerDied","Data":"df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36"} Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.566795 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" event={"ID":"58b0ae05-55dc-4238-9163-c8e06974ffca","Type":"ContainerDied","Data":"14de82752c3f80ce09390306e7e94d76d85a7ddcb61bc69d195c2a5bfdc7c861"} Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.566804 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.566869 4943 scope.go:117] "RemoveContainer" containerID="df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.601631 4943 generic.go:334] "Generic (PLEG): container finished" podID="8490c962-38c0-4be5-b940-d4c4d08c8a55" containerID="6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a" exitCode=0 Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.601683 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" event={"ID":"8490c962-38c0-4be5-b940-d4c4d08c8a55","Type":"ContainerDied","Data":"6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a"} Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.601713 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" event={"ID":"8490c962-38c0-4be5-b940-d4c4d08c8a55","Type":"ContainerDied","Data":"ef05790ae575ae735aad63f68e6efaae58dd78ad4799e8abbc7e3ebe1915b9b1"} Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.601772 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2g96c" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.621666 4943 scope.go:117] "RemoveContainer" containerID="df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36" Dec 04 10:21:09 crc kubenswrapper[4943]: E1204 10:21:09.622111 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36\": container with ID starting with df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36 not found: ID does not exist" containerID="df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.622155 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36"} err="failed to get container status \"df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36\": rpc error: code = NotFound desc = could not find container \"df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36\": container with ID starting with df57b5a358a5a6ab51b943e842e5497338503a57b67abf1bb1e2937fa9eb6b36 not found: ID does not exist" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.622176 4943 scope.go:117] "RemoveContainer" containerID="6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.635022 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2g96c"] Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.639226 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2g96c"] Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.643772 4943 scope.go:117] "RemoveContainer" containerID="6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a" Dec 04 10:21:09 crc kubenswrapper[4943]: E1204 10:21:09.645944 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a\": container with ID starting with 6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a not found: ID does not exist" containerID="6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.645993 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a"} err="failed to get container status \"6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a\": rpc error: code = NotFound desc = could not find container \"6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a\": container with ID starting with 6ebdaacc7d3c1df249e80f9512f3ccecc7de83f0183fed596ff10d662816819a not found: ID does not exist" Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.649755 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9"] Dec 04 10:21:09 crc kubenswrapper[4943]: I1204 10:21:09.654229 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-h9dk9"] Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.218886 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89"] Dec 04 10:21:10 crc kubenswrapper[4943]: E1204 10:21:10.219235 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.219266 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 10:21:10 crc kubenswrapper[4943]: E1204 10:21:10.219289 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8490c962-38c0-4be5-b940-d4c4d08c8a55" containerName="controller-manager" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.219296 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8490c962-38c0-4be5-b940-d4c4d08c8a55" containerName="controller-manager" Dec 04 10:21:10 crc kubenswrapper[4943]: E1204 10:21:10.219305 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58b0ae05-55dc-4238-9163-c8e06974ffca" containerName="route-controller-manager" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.219311 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="58b0ae05-55dc-4238-9163-c8e06974ffca" containerName="route-controller-manager" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.219466 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="58b0ae05-55dc-4238-9163-c8e06974ffca" containerName="route-controller-manager" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.219476 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="8490c962-38c0-4be5-b940-d4c4d08c8a55" containerName="controller-manager" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.219483 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.220476 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.222762 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.222762 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.223159 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.223261 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.223757 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.223904 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.226470 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-654c89ff58-n5rql"] Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.227322 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.229390 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.229886 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.229884 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.230051 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.230150 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.230245 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.234145 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89"] Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.236638 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.237861 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-654c89ff58-n5rql"] Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.323064 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrkqm\" (UniqueName: \"kubernetes.io/projected/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-kube-api-access-lrkqm\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.323146 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-client-ca\") pod \"route-controller-manager-fff48484f-9rz89\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.323303 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bztmh\" (UniqueName: \"kubernetes.io/projected/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-kube-api-access-bztmh\") pod \"route-controller-manager-fff48484f-9rz89\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.323355 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-serving-cert\") pod \"route-controller-manager-fff48484f-9rz89\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.323382 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-proxy-ca-bundles\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.323425 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-serving-cert\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.323454 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-config\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.323521 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-config\") pod \"route-controller-manager-fff48484f-9rz89\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.323572 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-client-ca\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.386472 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-654c89ff58-n5rql"] Dec 04 10:21:10 crc kubenswrapper[4943]: E1204 10:21:10.386988 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-lrkqm proxy-ca-bundles serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" podUID="8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.424695 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-client-ca\") pod \"route-controller-manager-fff48484f-9rz89\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.424772 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bztmh\" (UniqueName: \"kubernetes.io/projected/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-kube-api-access-bztmh\") pod \"route-controller-manager-fff48484f-9rz89\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.424812 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-serving-cert\") pod \"route-controller-manager-fff48484f-9rz89\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.424838 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-proxy-ca-bundles\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.425612 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-serving-cert\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.425651 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-config\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.425719 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-config\") pod \"route-controller-manager-fff48484f-9rz89\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.425782 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-client-ca\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.425822 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrkqm\" (UniqueName: \"kubernetes.io/projected/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-kube-api-access-lrkqm\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.426539 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-client-ca\") pod \"route-controller-manager-fff48484f-9rz89\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.426609 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-proxy-ca-bundles\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.426783 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-client-ca\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.427893 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-config\") pod \"route-controller-manager-fff48484f-9rz89\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.427998 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-config\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.429337 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89"] Dec 04 10:21:10 crc kubenswrapper[4943]: E1204 10:21:10.429724 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-bztmh serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" podUID="8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.431285 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-serving-cert\") pod \"route-controller-manager-fff48484f-9rz89\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.431835 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-serving-cert\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.453268 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bztmh\" (UniqueName: \"kubernetes.io/projected/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-kube-api-access-bztmh\") pod \"route-controller-manager-fff48484f-9rz89\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.461917 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrkqm\" (UniqueName: \"kubernetes.io/projected/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-kube-api-access-lrkqm\") pod \"controller-manager-654c89ff58-n5rql\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.574277 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58b0ae05-55dc-4238-9163-c8e06974ffca" path="/var/lib/kubelet/pods/58b0ae05-55dc-4238-9163-c8e06974ffca/volumes" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.575777 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8490c962-38c0-4be5-b940-d4c4d08c8a55" path="/var/lib/kubelet/pods/8490c962-38c0-4be5-b940-d4c4d08c8a55/volumes" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.611391 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.611435 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.624504 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.627927 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bztmh\" (UniqueName: \"kubernetes.io/projected/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-kube-api-access-bztmh\") pod \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.628008 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-serving-cert\") pod \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.628107 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-config\") pod \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.628133 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-client-ca\") pod \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\" (UID: \"8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7\") " Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.628696 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-config" (OuterVolumeSpecName: "config") pod "8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7" (UID: "8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.629179 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-client-ca" (OuterVolumeSpecName: "client-ca") pod "8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7" (UID: "8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.631616 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-kube-api-access-bztmh" (OuterVolumeSpecName: "kube-api-access-bztmh") pod "8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7" (UID: "8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7"). InnerVolumeSpecName "kube-api-access-bztmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.633091 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7" (UID: "8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.633782 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.729867 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-serving-cert\") pod \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.730014 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-proxy-ca-bundles\") pod \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.730066 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-client-ca\") pod \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.730091 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-config\") pod \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.730158 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrkqm\" (UniqueName: \"kubernetes.io/projected/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-kube-api-access-lrkqm\") pod \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\" (UID: \"8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa\") " Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.730435 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.730464 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.730477 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bztmh\" (UniqueName: \"kubernetes.io/projected/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-kube-api-access-bztmh\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.730492 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.730741 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-client-ca" (OuterVolumeSpecName: "client-ca") pod "8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa" (UID: "8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.730777 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa" (UID: "8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.730966 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-config" (OuterVolumeSpecName: "config") pod "8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa" (UID: "8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.733035 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa" (UID: "8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.733115 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-kube-api-access-lrkqm" (OuterVolumeSpecName: "kube-api-access-lrkqm") pod "8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa" (UID: "8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa"). InnerVolumeSpecName "kube-api-access-lrkqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.831947 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.831983 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.831998 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.832008 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:10 crc kubenswrapper[4943]: I1204 10:21:10.832020 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrkqm\" (UniqueName: \"kubernetes.io/projected/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa-kube-api-access-lrkqm\") on node \"crc\" DevicePath \"\"" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.615794 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.615812 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-654c89ff58-n5rql" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.659171 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-76b976544-pfp2x"] Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.660220 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.663370 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-654c89ff58-n5rql"] Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.663548 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.663616 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.663799 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.663940 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.664053 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.664153 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.669782 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.671014 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-654c89ff58-n5rql"] Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.675847 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76b976544-pfp2x"] Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.715442 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89"] Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.721343 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fff48484f-9rz89"] Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.744286 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-config\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.744340 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-client-ca\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.744359 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-proxy-ca-bundles\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.744452 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87d9564-2875-4550-af91-a9257d9ae83c-serving-cert\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.744627 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdwb6\" (UniqueName: \"kubernetes.io/projected/a87d9564-2875-4550-af91-a9257d9ae83c-kube-api-access-wdwb6\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.846084 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-config\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.846136 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-client-ca\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.846154 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-proxy-ca-bundles\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.846177 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87d9564-2875-4550-af91-a9257d9ae83c-serving-cert\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.846237 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdwb6\" (UniqueName: \"kubernetes.io/projected/a87d9564-2875-4550-af91-a9257d9ae83c-kube-api-access-wdwb6\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.847737 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-client-ca\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.849228 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-config\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.850537 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-proxy-ca-bundles\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.851735 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87d9564-2875-4550-af91-a9257d9ae83c-serving-cert\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.863022 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdwb6\" (UniqueName: \"kubernetes.io/projected/a87d9564-2875-4550-af91-a9257d9ae83c-kube-api-access-wdwb6\") pod \"controller-manager-76b976544-pfp2x\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:11 crc kubenswrapper[4943]: I1204 10:21:11.997688 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:12 crc kubenswrapper[4943]: I1204 10:21:12.206646 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76b976544-pfp2x"] Dec 04 10:21:12 crc kubenswrapper[4943]: W1204 10:21:12.212728 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda87d9564_2875_4550_af91_a9257d9ae83c.slice/crio-705755991edfca6c90c14731e15929cbc113ee14eaca05f42170171fafb9d3f8 WatchSource:0}: Error finding container 705755991edfca6c90c14731e15929cbc113ee14eaca05f42170171fafb9d3f8: Status 404 returned error can't find the container with id 705755991edfca6c90c14731e15929cbc113ee14eaca05f42170171fafb9d3f8 Dec 04 10:21:12 crc kubenswrapper[4943]: I1204 10:21:12.572698 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7" path="/var/lib/kubelet/pods/8a818fce-d2e5-42cd-a5aa-fc9d4e65e8e7/volumes" Dec 04 10:21:12 crc kubenswrapper[4943]: I1204 10:21:12.573163 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa" path="/var/lib/kubelet/pods/8febdf5a-8fe0-4dcc-8a9a-66d9e3e22daa/volumes" Dec 04 10:21:12 crc kubenswrapper[4943]: I1204 10:21:12.622818 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" event={"ID":"a87d9564-2875-4550-af91-a9257d9ae83c","Type":"ContainerStarted","Data":"52a18460e2920491a24acab6877cca30f8730bf4a2260abe4aa1754e8cbffde6"} Dec 04 10:21:12 crc kubenswrapper[4943]: I1204 10:21:12.623139 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" event={"ID":"a87d9564-2875-4550-af91-a9257d9ae83c","Type":"ContainerStarted","Data":"705755991edfca6c90c14731e15929cbc113ee14eaca05f42170171fafb9d3f8"} Dec 04 10:21:12 crc kubenswrapper[4943]: I1204 10:21:12.623797 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:12 crc kubenswrapper[4943]: I1204 10:21:12.645187 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:21:12 crc kubenswrapper[4943]: I1204 10:21:12.662884 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" podStartSLOduration=2.662849849 podStartE2EDuration="2.662849849s" podCreationTimestamp="2025-12-04 10:21:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:21:12.659918696 +0000 UTC m=+321.248894564" watchObservedRunningTime="2025-12-04 10:21:12.662849849 +0000 UTC m=+321.251825717" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.220739 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx"] Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.221444 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.224515 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.224547 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.224818 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.224922 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.225179 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.225390 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.240893 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx"] Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.335915 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clhnw\" (UniqueName: \"kubernetes.io/projected/1922c24e-5702-4d9a-a5b9-2abaa98871f4-kube-api-access-clhnw\") pod \"route-controller-manager-8d8c54746-hkqbx\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.336167 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1922c24e-5702-4d9a-a5b9-2abaa98871f4-client-ca\") pod \"route-controller-manager-8d8c54746-hkqbx\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.336257 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1922c24e-5702-4d9a-a5b9-2abaa98871f4-config\") pod \"route-controller-manager-8d8c54746-hkqbx\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.336416 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1922c24e-5702-4d9a-a5b9-2abaa98871f4-serving-cert\") pod \"route-controller-manager-8d8c54746-hkqbx\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.437048 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1922c24e-5702-4d9a-a5b9-2abaa98871f4-serving-cert\") pod \"route-controller-manager-8d8c54746-hkqbx\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.437134 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clhnw\" (UniqueName: \"kubernetes.io/projected/1922c24e-5702-4d9a-a5b9-2abaa98871f4-kube-api-access-clhnw\") pod \"route-controller-manager-8d8c54746-hkqbx\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.437184 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1922c24e-5702-4d9a-a5b9-2abaa98871f4-client-ca\") pod \"route-controller-manager-8d8c54746-hkqbx\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.437264 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1922c24e-5702-4d9a-a5b9-2abaa98871f4-config\") pod \"route-controller-manager-8d8c54746-hkqbx\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.438263 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1922c24e-5702-4d9a-a5b9-2abaa98871f4-client-ca\") pod \"route-controller-manager-8d8c54746-hkqbx\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.438466 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1922c24e-5702-4d9a-a5b9-2abaa98871f4-config\") pod \"route-controller-manager-8d8c54746-hkqbx\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.442375 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1922c24e-5702-4d9a-a5b9-2abaa98871f4-serving-cert\") pod \"route-controller-manager-8d8c54746-hkqbx\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.456437 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clhnw\" (UniqueName: \"kubernetes.io/projected/1922c24e-5702-4d9a-a5b9-2abaa98871f4-kube-api-access-clhnw\") pod \"route-controller-manager-8d8c54746-hkqbx\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.539626 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:14 crc kubenswrapper[4943]: I1204 10:21:14.734268 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx"] Dec 04 10:21:15 crc kubenswrapper[4943]: I1204 10:21:15.644420 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" event={"ID":"1922c24e-5702-4d9a-a5b9-2abaa98871f4","Type":"ContainerStarted","Data":"a1145e8481ac2d807a25e1e8948571030bd5398f4b6e7d7ea4d5018e7f83cdaf"} Dec 04 10:21:15 crc kubenswrapper[4943]: I1204 10:21:15.644796 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" event={"ID":"1922c24e-5702-4d9a-a5b9-2abaa98871f4","Type":"ContainerStarted","Data":"c49aff73f6d86281a4db981c1529245153182284e9951e9f222dfadbf36d39d4"} Dec 04 10:21:15 crc kubenswrapper[4943]: I1204 10:21:15.645332 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:15 crc kubenswrapper[4943]: I1204 10:21:15.670089 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" podStartSLOduration=5.670036987 podStartE2EDuration="5.670036987s" podCreationTimestamp="2025-12-04 10:21:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:21:15.662952076 +0000 UTC m=+324.251927944" watchObservedRunningTime="2025-12-04 10:21:15.670036987 +0000 UTC m=+324.259012855" Dec 04 10:21:15 crc kubenswrapper[4943]: I1204 10:21:15.817623 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.188168 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-ttflv"] Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.189387 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.196759 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-ttflv"] Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.321095 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-registry-certificates\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.321185 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-trusted-ca\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.321272 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-bound-sa-token\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.321300 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-registry-tls\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.321319 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.321339 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxk4h\" (UniqueName: \"kubernetes.io/projected/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-kube-api-access-nxk4h\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.321372 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.321394 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.342167 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.422284 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-bound-sa-token\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.422586 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-registry-tls\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.422686 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.422777 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxk4h\" (UniqueName: \"kubernetes.io/projected/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-kube-api-access-nxk4h\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.422900 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.423007 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-registry-certificates\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.423128 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-trusted-ca\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.423870 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-ca-trust-extracted\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.424413 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-trusted-ca\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.424553 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-registry-certificates\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.430032 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-registry-tls\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.430311 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-installation-pull-secrets\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.439979 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxk4h\" (UniqueName: \"kubernetes.io/projected/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-kube-api-access-nxk4h\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.440703 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8-bound-sa-token\") pod \"image-registry-66df7c8f76-ttflv\" (UID: \"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8\") " pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.511018 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:34 crc kubenswrapper[4943]: I1204 10:21:34.925290 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-ttflv"] Dec 04 10:21:35 crc kubenswrapper[4943]: I1204 10:21:35.758024 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" event={"ID":"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8","Type":"ContainerStarted","Data":"1d48751bf55dd978bc0af53acd2fa7e6b46c483d90df588566fdf962320dc64e"} Dec 04 10:21:35 crc kubenswrapper[4943]: I1204 10:21:35.758411 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" event={"ID":"2b15ae7f-a4e4-4f9d-96f3-3af589baf7f8","Type":"ContainerStarted","Data":"14d1bc93eb17fda06f08d904809c7e8c858f72bc05a3e03cdf57fa6b29bcc973"} Dec 04 10:21:35 crc kubenswrapper[4943]: I1204 10:21:35.758434 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:35 crc kubenswrapper[4943]: I1204 10:21:35.776188 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" podStartSLOduration=1.776165517 podStartE2EDuration="1.776165517s" podCreationTimestamp="2025-12-04 10:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:21:35.774280973 +0000 UTC m=+344.363256851" watchObservedRunningTime="2025-12-04 10:21:35.776165517 +0000 UTC m=+344.365141385" Dec 04 10:21:48 crc kubenswrapper[4943]: I1204 10:21:48.628836 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:21:48 crc kubenswrapper[4943]: I1204 10:21:48.629401 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:21:54 crc kubenswrapper[4943]: I1204 10:21:54.516422 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-ttflv" Dec 04 10:21:54 crc kubenswrapper[4943]: I1204 10:21:54.576579 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wmxwg"] Dec 04 10:22:18 crc kubenswrapper[4943]: I1204 10:22:18.629672 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:22:18 crc kubenswrapper[4943]: I1204 10:22:18.630378 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.332847 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l7p8s"] Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.333339 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l7p8s" podUID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" containerName="registry-server" containerID="cri-o://5d7a4fb66febdd49aca0d0f99bb524bb59009d29c5081a886304e5ee3a455c6f" gracePeriod=30 Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.344477 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jcsk7"] Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.344815 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jcsk7" podUID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" containerName="registry-server" containerID="cri-o://284bb10b585888ced45051d724242bdb9424f39fd99c729021086fbb344cf1a4" gracePeriod=30 Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.349772 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ph88v"] Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.350024 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" podUID="3e183218-08a1-4890-a6af-d56f3c7db2be" containerName="marketplace-operator" containerID="cri-o://698a6d0e6901b22f5d4757974553cd340604d29a3013380b7eb9ce2eae17e20e" gracePeriod=30 Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.381906 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9rd8"] Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.382327 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q9rd8" podUID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" containerName="registry-server" containerID="cri-o://917cf408413d788ff9a009ad32f4722959d1d70a5ce5478a79910ce5ea47277d" gracePeriod=30 Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.398108 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bwkhz"] Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.399289 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.401792 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gtr9n"] Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.402099 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gtr9n" podUID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" containerName="registry-server" containerID="cri-o://b3a7f0acb34017d8a70596513a6daee081b574ee93b80811df18c457c996ae70" gracePeriod=30 Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.416895 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bwkhz"] Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.572790 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jk9h\" (UniqueName: \"kubernetes.io/projected/04c5f3f6-69da-4c89-801f-5e3d519c80ae-kube-api-access-7jk9h\") pod \"marketplace-operator-79b997595-bwkhz\" (UID: \"04c5f3f6-69da-4c89-801f-5e3d519c80ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.572892 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/04c5f3f6-69da-4c89-801f-5e3d519c80ae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bwkhz\" (UID: \"04c5f3f6-69da-4c89-801f-5e3d519c80ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.573021 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/04c5f3f6-69da-4c89-801f-5e3d519c80ae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bwkhz\" (UID: \"04c5f3f6-69da-4c89-801f-5e3d519c80ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.630365 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" podUID="a6da7528-9b01-40b4-88d3-e78a3c178300" containerName="registry" containerID="cri-o://839ccfda6aed59bced160eea597855bf4711831b79485a05de7c9d5d51274930" gracePeriod=30 Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.674108 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/04c5f3f6-69da-4c89-801f-5e3d519c80ae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bwkhz\" (UID: \"04c5f3f6-69da-4c89-801f-5e3d519c80ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.674179 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/04c5f3f6-69da-4c89-801f-5e3d519c80ae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bwkhz\" (UID: \"04c5f3f6-69da-4c89-801f-5e3d519c80ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.674285 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jk9h\" (UniqueName: \"kubernetes.io/projected/04c5f3f6-69da-4c89-801f-5e3d519c80ae-kube-api-access-7jk9h\") pod \"marketplace-operator-79b997595-bwkhz\" (UID: \"04c5f3f6-69da-4c89-801f-5e3d519c80ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.675903 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/04c5f3f6-69da-4c89-801f-5e3d519c80ae-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bwkhz\" (UID: \"04c5f3f6-69da-4c89-801f-5e3d519c80ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.681020 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/04c5f3f6-69da-4c89-801f-5e3d519c80ae-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bwkhz\" (UID: \"04c5f3f6-69da-4c89-801f-5e3d519c80ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.692764 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jk9h\" (UniqueName: \"kubernetes.io/projected/04c5f3f6-69da-4c89-801f-5e3d519c80ae-kube-api-access-7jk9h\") pod \"marketplace-operator-79b997595-bwkhz\" (UID: \"04c5f3f6-69da-4c89-801f-5e3d519c80ae\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:19 crc kubenswrapper[4943]: I1204 10:22:19.721395 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.047923 4943 generic.go:334] "Generic (PLEG): container finished" podID="a6da7528-9b01-40b4-88d3-e78a3c178300" containerID="839ccfda6aed59bced160eea597855bf4711831b79485a05de7c9d5d51274930" exitCode=0 Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.047989 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" event={"ID":"a6da7528-9b01-40b4-88d3-e78a3c178300","Type":"ContainerDied","Data":"839ccfda6aed59bced160eea597855bf4711831b79485a05de7c9d5d51274930"} Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.050486 4943 generic.go:334] "Generic (PLEG): container finished" podID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" containerID="b3a7f0acb34017d8a70596513a6daee081b574ee93b80811df18c457c996ae70" exitCode=0 Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.050554 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtr9n" event={"ID":"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa","Type":"ContainerDied","Data":"b3a7f0acb34017d8a70596513a6daee081b574ee93b80811df18c457c996ae70"} Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.052613 4943 generic.go:334] "Generic (PLEG): container finished" podID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" containerID="284bb10b585888ced45051d724242bdb9424f39fd99c729021086fbb344cf1a4" exitCode=0 Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.052666 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jcsk7" event={"ID":"d22ad643-0a77-49fb-a7f1-54603fe0f82d","Type":"ContainerDied","Data":"284bb10b585888ced45051d724242bdb9424f39fd99c729021086fbb344cf1a4"} Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.054505 4943 generic.go:334] "Generic (PLEG): container finished" podID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" containerID="917cf408413d788ff9a009ad32f4722959d1d70a5ce5478a79910ce5ea47277d" exitCode=0 Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.054568 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9rd8" event={"ID":"2c9ccdf7-acc0-481a-b1de-5172e16ef22c","Type":"ContainerDied","Data":"917cf408413d788ff9a009ad32f4722959d1d70a5ce5478a79910ce5ea47277d"} Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.056367 4943 generic.go:334] "Generic (PLEG): container finished" podID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" containerID="5d7a4fb66febdd49aca0d0f99bb524bb59009d29c5081a886304e5ee3a455c6f" exitCode=0 Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.056427 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7p8s" event={"ID":"0d555ec7-2800-45a7-a1a2-27b1a435b0d4","Type":"ContainerDied","Data":"5d7a4fb66febdd49aca0d0f99bb524bb59009d29c5081a886304e5ee3a455c6f"} Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.057783 4943 generic.go:334] "Generic (PLEG): container finished" podID="3e183218-08a1-4890-a6af-d56f3c7db2be" containerID="698a6d0e6901b22f5d4757974553cd340604d29a3013380b7eb9ce2eae17e20e" exitCode=0 Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.057815 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" event={"ID":"3e183218-08a1-4890-a6af-d56f3c7db2be","Type":"ContainerDied","Data":"698a6d0e6901b22f5d4757974553cd340604d29a3013380b7eb9ce2eae17e20e"} Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.179708 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bwkhz"] Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.421176 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.445176 4943 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-ph88v container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.445270 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" podUID="3e183218-08a1-4890-a6af-d56f3c7db2be" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.501971 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwhw9\" (UniqueName: \"kubernetes.io/projected/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-kube-api-access-gwhw9\") pod \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\" (UID: \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.502038 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-utilities\") pod \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\" (UID: \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.502142 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-catalog-content\") pod \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\" (UID: \"0d555ec7-2800-45a7-a1a2-27b1a435b0d4\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.503019 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-utilities" (OuterVolumeSpecName: "utilities") pod "0d555ec7-2800-45a7-a1a2-27b1a435b0d4" (UID: "0d555ec7-2800-45a7-a1a2-27b1a435b0d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.508867 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-kube-api-access-gwhw9" (OuterVolumeSpecName: "kube-api-access-gwhw9") pod "0d555ec7-2800-45a7-a1a2-27b1a435b0d4" (UID: "0d555ec7-2800-45a7-a1a2-27b1a435b0d4"). InnerVolumeSpecName "kube-api-access-gwhw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.569739 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d555ec7-2800-45a7-a1a2-27b1a435b0d4" (UID: "0d555ec7-2800-45a7-a1a2-27b1a435b0d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.592063 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.603377 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwhw9\" (UniqueName: \"kubernetes.io/projected/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-kube-api-access-gwhw9\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.603412 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.603423 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d555ec7-2800-45a7-a1a2-27b1a435b0d4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.617316 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.704114 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-bound-sa-token\") pod \"a6da7528-9b01-40b4-88d3-e78a3c178300\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.705355 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"a6da7528-9b01-40b4-88d3-e78a3c178300\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.705868 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-registry-tls\") pod \"a6da7528-9b01-40b4-88d3-e78a3c178300\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.706353 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a6da7528-9b01-40b4-88d3-e78a3c178300-ca-trust-extracted\") pod \"a6da7528-9b01-40b4-88d3-e78a3c178300\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.706489 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d22ad643-0a77-49fb-a7f1-54603fe0f82d-utilities\") pod \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\" (UID: \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.707634 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6da7528-9b01-40b4-88d3-e78a3c178300-trusted-ca\") pod \"a6da7528-9b01-40b4-88d3-e78a3c178300\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.707869 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a6da7528-9b01-40b4-88d3-e78a3c178300-registry-certificates\") pod \"a6da7528-9b01-40b4-88d3-e78a3c178300\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.707993 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d22ad643-0a77-49fb-a7f1-54603fe0f82d-catalog-content\") pod \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\" (UID: \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.708218 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfjpc\" (UniqueName: \"kubernetes.io/projected/d22ad643-0a77-49fb-a7f1-54603fe0f82d-kube-api-access-rfjpc\") pod \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\" (UID: \"d22ad643-0a77-49fb-a7f1-54603fe0f82d\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.708344 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a6da7528-9b01-40b4-88d3-e78a3c178300-installation-pull-secrets\") pod \"a6da7528-9b01-40b4-88d3-e78a3c178300\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.708487 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxxrv\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-kube-api-access-sxxrv\") pod \"a6da7528-9b01-40b4-88d3-e78a3c178300\" (UID: \"a6da7528-9b01-40b4-88d3-e78a3c178300\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.709976 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d22ad643-0a77-49fb-a7f1-54603fe0f82d-utilities" (OuterVolumeSpecName: "utilities") pod "d22ad643-0a77-49fb-a7f1-54603fe0f82d" (UID: "d22ad643-0a77-49fb-a7f1-54603fe0f82d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.714568 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a6da7528-9b01-40b4-88d3-e78a3c178300" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.714868 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6da7528-9b01-40b4-88d3-e78a3c178300-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "a6da7528-9b01-40b4-88d3-e78a3c178300" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.715147 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6da7528-9b01-40b4-88d3-e78a3c178300-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a6da7528-9b01-40b4-88d3-e78a3c178300" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.716938 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-kube-api-access-sxxrv" (OuterVolumeSpecName: "kube-api-access-sxxrv") pod "a6da7528-9b01-40b4-88d3-e78a3c178300" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300"). InnerVolumeSpecName "kube-api-access-sxxrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.717820 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6da7528-9b01-40b4-88d3-e78a3c178300-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "a6da7528-9b01-40b4-88d3-e78a3c178300" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.719523 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "a6da7528-9b01-40b4-88d3-e78a3c178300" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.721326 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.726770 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "a6da7528-9b01-40b4-88d3-e78a3c178300" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.730794 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d22ad643-0a77-49fb-a7f1-54603fe0f82d-kube-api-access-rfjpc" (OuterVolumeSpecName: "kube-api-access-rfjpc") pod "d22ad643-0a77-49fb-a7f1-54603fe0f82d" (UID: "d22ad643-0a77-49fb-a7f1-54603fe0f82d"). InnerVolumeSpecName "kube-api-access-rfjpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.735082 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6da7528-9b01-40b4-88d3-e78a3c178300-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "a6da7528-9b01-40b4-88d3-e78a3c178300" (UID: "a6da7528-9b01-40b4-88d3-e78a3c178300"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.738382 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.752176 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.785730 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d22ad643-0a77-49fb-a7f1-54603fe0f82d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d22ad643-0a77-49fb-a7f1-54603fe0f82d" (UID: "d22ad643-0a77-49fb-a7f1-54603fe0f82d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.810602 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3e183218-08a1-4890-a6af-d56f3c7db2be-marketplace-operator-metrics\") pod \"3e183218-08a1-4890-a6af-d56f3c7db2be\" (UID: \"3e183218-08a1-4890-a6af-d56f3c7db2be\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.810715 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-utilities\") pod \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\" (UID: \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.810748 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z96bh\" (UniqueName: \"kubernetes.io/projected/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-kube-api-access-z96bh\") pod \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\" (UID: \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.810802 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-utilities\") pod \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\" (UID: \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.810838 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mvp7\" (UniqueName: \"kubernetes.io/projected/3e183218-08a1-4890-a6af-d56f3c7db2be-kube-api-access-8mvp7\") pod \"3e183218-08a1-4890-a6af-d56f3c7db2be\" (UID: \"3e183218-08a1-4890-a6af-d56f3c7db2be\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.810867 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99lx2\" (UniqueName: \"kubernetes.io/projected/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-kube-api-access-99lx2\") pod \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\" (UID: \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.810921 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-catalog-content\") pod \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\" (UID: \"2c9ccdf7-acc0-481a-b1de-5172e16ef22c\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.810942 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-catalog-content\") pod \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\" (UID: \"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.810971 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e183218-08a1-4890-a6af-d56f3c7db2be-marketplace-trusted-ca\") pod \"3e183218-08a1-4890-a6af-d56f3c7db2be\" (UID: \"3e183218-08a1-4890-a6af-d56f3c7db2be\") " Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.811861 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-utilities" (OuterVolumeSpecName: "utilities") pod "2c9ccdf7-acc0-481a-b1de-5172e16ef22c" (UID: "2c9ccdf7-acc0-481a-b1de-5172e16ef22c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.814974 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e183218-08a1-4890-a6af-d56f3c7db2be-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "3e183218-08a1-4890-a6af-d56f3c7db2be" (UID: "3e183218-08a1-4890-a6af-d56f3c7db2be"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.815346 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-utilities" (OuterVolumeSpecName: "utilities") pod "3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" (UID: "3a3b7592-0bdb-4655-87a3-86ecacc6a0fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.815709 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-kube-api-access-99lx2" (OuterVolumeSpecName: "kube-api-access-99lx2") pod "2c9ccdf7-acc0-481a-b1de-5172e16ef22c" (UID: "2c9ccdf7-acc0-481a-b1de-5172e16ef22c"). InnerVolumeSpecName "kube-api-access-99lx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.816763 4943 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a6da7528-9b01-40b4-88d3-e78a3c178300-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.816797 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d22ad643-0a77-49fb-a7f1-54603fe0f82d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.816812 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfjpc\" (UniqueName: \"kubernetes.io/projected/d22ad643-0a77-49fb-a7f1-54603fe0f82d-kube-api-access-rfjpc\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.816828 4943 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a6da7528-9b01-40b4-88d3-e78a3c178300-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.816841 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxxrv\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-kube-api-access-sxxrv\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.817558 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e183218-08a1-4890-a6af-d56f3c7db2be-kube-api-access-8mvp7" (OuterVolumeSpecName: "kube-api-access-8mvp7") pod "3e183218-08a1-4890-a6af-d56f3c7db2be" (UID: "3e183218-08a1-4890-a6af-d56f3c7db2be"). InnerVolumeSpecName "kube-api-access-8mvp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.819320 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.819371 4943 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.819389 4943 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a6da7528-9b01-40b4-88d3-e78a3c178300-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.819400 4943 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a6da7528-9b01-40b4-88d3-e78a3c178300-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.819422 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d22ad643-0a77-49fb-a7f1-54603fe0f82d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.819431 4943 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6da7528-9b01-40b4-88d3-e78a3c178300-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.820423 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-kube-api-access-z96bh" (OuterVolumeSpecName: "kube-api-access-z96bh") pod "3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" (UID: "3a3b7592-0bdb-4655-87a3-86ecacc6a0fa"). InnerVolumeSpecName "kube-api-access-z96bh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.833968 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e183218-08a1-4890-a6af-d56f3c7db2be-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "3e183218-08a1-4890-a6af-d56f3c7db2be" (UID: "3e183218-08a1-4890-a6af-d56f3c7db2be"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.834998 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c9ccdf7-acc0-481a-b1de-5172e16ef22c" (UID: "2c9ccdf7-acc0-481a-b1de-5172e16ef22c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.920544 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99lx2\" (UniqueName: \"kubernetes.io/projected/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-kube-api-access-99lx2\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.920586 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c9ccdf7-acc0-481a-b1de-5172e16ef22c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.920599 4943 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e183218-08a1-4890-a6af-d56f3c7db2be-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.920611 4943 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3e183218-08a1-4890-a6af-d56f3c7db2be-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.920625 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.920637 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z96bh\" (UniqueName: \"kubernetes.io/projected/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-kube-api-access-z96bh\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.920648 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mvp7\" (UniqueName: \"kubernetes.io/projected/3e183218-08a1-4890-a6af-d56f3c7db2be-kube-api-access-8mvp7\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:20 crc kubenswrapper[4943]: I1204 10:22:20.933847 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" (UID: "3a3b7592-0bdb-4655-87a3-86ecacc6a0fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.022125 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.063916 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" event={"ID":"a6da7528-9b01-40b4-88d3-e78a3c178300","Type":"ContainerDied","Data":"c3e3832740fa9bee9196c6115d250312879a88b42171b0e5d88dc4f82fb57c8a"} Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.063939 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wmxwg" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.064031 4943 scope.go:117] "RemoveContainer" containerID="839ccfda6aed59bced160eea597855bf4711831b79485a05de7c9d5d51274930" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.071216 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gtr9n" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.071184 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gtr9n" event={"ID":"3a3b7592-0bdb-4655-87a3-86ecacc6a0fa","Type":"ContainerDied","Data":"8e5d4f53edddcb2cc65345650130a916767d1b7aa386111ef26c5225aa7ea1af"} Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.073991 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jcsk7" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.073990 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jcsk7" event={"ID":"d22ad643-0a77-49fb-a7f1-54603fe0f82d","Type":"ContainerDied","Data":"1c52d974a54ca301b60ff950116233e35c8fcfb42112157e236698a1cc91b7dd"} Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.077041 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q9rd8" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.077061 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9rd8" event={"ID":"2c9ccdf7-acc0-481a-b1de-5172e16ef22c","Type":"ContainerDied","Data":"775938d2763cb2e750e2368545b222f05d565163d9f7ff3a60c338d5eda7bae1"} Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.084425 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" event={"ID":"04c5f3f6-69da-4c89-801f-5e3d519c80ae","Type":"ContainerStarted","Data":"b321aae590031f38455c2b4c644afdacd31a3089e060e9ab8b60912a631f9389"} Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.084507 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" event={"ID":"04c5f3f6-69da-4c89-801f-5e3d519c80ae","Type":"ContainerStarted","Data":"b5ff0342674450e114e2c3c397a8c55afd46b7d2e1e6967a86978264e78f57d8"} Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.084933 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.097533 4943 scope.go:117] "RemoveContainer" containerID="b3a7f0acb34017d8a70596513a6daee081b574ee93b80811df18c457c996ae70" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.100825 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l7p8s" event={"ID":"0d555ec7-2800-45a7-a1a2-27b1a435b0d4","Type":"ContainerDied","Data":"3cdac60ea1b14a02105a384481bc8c94b501b757ee13247fa16dae1703d0de2a"} Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.100910 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l7p8s" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.112476 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" event={"ID":"3e183218-08a1-4890-a6af-d56f3c7db2be","Type":"ContainerDied","Data":"2fbb30276478cfd9c815dcae8d1122ebe4e49a5e5d750366871e0793a3eeb28a"} Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.112579 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ph88v" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.118939 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wmxwg"] Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.132325 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wmxwg"] Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.132835 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.133666 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bwkhz" podStartSLOduration=2.133628708 podStartE2EDuration="2.133628708s" podCreationTimestamp="2025-12-04 10:22:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:22:21.125276754 +0000 UTC m=+389.714252622" watchObservedRunningTime="2025-12-04 10:22:21.133628708 +0000 UTC m=+389.722604586" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.133702 4943 scope.go:117] "RemoveContainer" containerID="a826e9bae054356100ca51c7ea5adc95f02d735ac4ba137ebdbf0b4987098016" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.154170 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gtr9n"] Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.160548 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gtr9n"] Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.167767 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l7p8s"] Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.170976 4943 scope.go:117] "RemoveContainer" containerID="98b4197117fbc4df73e705d92977317e05478527b4efefc0e9e63f3f26fdaa47" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.172359 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l7p8s"] Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.179800 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jcsk7"] Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.182067 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jcsk7"] Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.188367 4943 scope.go:117] "RemoveContainer" containerID="284bb10b585888ced45051d724242bdb9424f39fd99c729021086fbb344cf1a4" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.189400 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ph88v"] Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.194721 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ph88v"] Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.199773 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9rd8"] Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.204959 4943 scope.go:117] "RemoveContainer" containerID="12da88b3bc918fe777d34adea50a0b4eb3b7b7379221e26f65414fc9b5e59071" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.215809 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9rd8"] Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.234874 4943 scope.go:117] "RemoveContainer" containerID="f87bea5167ff7b26158aabb080b78d71611013b8e5bbe5d1109a4ae8f49c22cf" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.252154 4943 scope.go:117] "RemoveContainer" containerID="917cf408413d788ff9a009ad32f4722959d1d70a5ce5478a79910ce5ea47277d" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.265648 4943 scope.go:117] "RemoveContainer" containerID="7d856df5c5c7a8fc2ad0d6654f66e626ab449b327f45a5e611e60fa0c0b1fb1c" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.283717 4943 scope.go:117] "RemoveContainer" containerID="356985a1da476424d3b644ea0bffaba449408ee868825f7e6734bf11c142f6b0" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.298128 4943 scope.go:117] "RemoveContainer" containerID="5d7a4fb66febdd49aca0d0f99bb524bb59009d29c5081a886304e5ee3a455c6f" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.314478 4943 scope.go:117] "RemoveContainer" containerID="b49d16cea7fdbb8aaa63a329bf68f393dd91587a2c440d2875542cb1d80c58c2" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.330848 4943 scope.go:117] "RemoveContainer" containerID="5097a1826504d7701f463452245f87b1b01b42e6f863f0fea7f06d282852d7f1" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.345126 4943 scope.go:117] "RemoveContainer" containerID="698a6d0e6901b22f5d4757974553cd340604d29a3013380b7eb9ce2eae17e20e" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.957499 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c44sb"] Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958101 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" containerName="extract-utilities" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958133 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" containerName="extract-utilities" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958148 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" containerName="extract-content" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958156 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" containerName="extract-content" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958167 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" containerName="registry-server" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958175 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" containerName="registry-server" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958188 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" containerName="extract-utilities" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958196 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" containerName="extract-utilities" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958222 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" containerName="extract-content" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958229 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" containerName="extract-content" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958244 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" containerName="extract-content" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958251 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" containerName="extract-content" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958261 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" containerName="extract-utilities" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958268 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" containerName="extract-utilities" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958277 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" containerName="extract-utilities" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958285 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" containerName="extract-utilities" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958295 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" containerName="registry-server" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958302 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" containerName="registry-server" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958313 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" containerName="registry-server" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958320 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" containerName="registry-server" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958371 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" containerName="registry-server" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958380 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" containerName="registry-server" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958392 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e183218-08a1-4890-a6af-d56f3c7db2be" containerName="marketplace-operator" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958399 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e183218-08a1-4890-a6af-d56f3c7db2be" containerName="marketplace-operator" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958411 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" containerName="extract-content" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958419 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" containerName="extract-content" Dec 04 10:22:21 crc kubenswrapper[4943]: E1204 10:22:21.958429 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6da7528-9b01-40b4-88d3-e78a3c178300" containerName="registry" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958436 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6da7528-9b01-40b4-88d3-e78a3c178300" containerName="registry" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958588 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" containerName="registry-server" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958601 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" containerName="registry-server" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958611 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6da7528-9b01-40b4-88d3-e78a3c178300" containerName="registry" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958625 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" containerName="registry-server" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958635 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" containerName="registry-server" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.958645 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e183218-08a1-4890-a6af-d56f3c7db2be" containerName="marketplace-operator" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.959695 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.962599 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 10:22:21 crc kubenswrapper[4943]: I1204 10:22:21.991265 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c44sb"] Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.037221 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15b454e0-b590-41a9-b496-bf6577afbfed-utilities\") pod \"redhat-marketplace-c44sb\" (UID: \"15b454e0-b590-41a9-b496-bf6577afbfed\") " pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.037407 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15b454e0-b590-41a9-b496-bf6577afbfed-catalog-content\") pod \"redhat-marketplace-c44sb\" (UID: \"15b454e0-b590-41a9-b496-bf6577afbfed\") " pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.037561 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hljr4\" (UniqueName: \"kubernetes.io/projected/15b454e0-b590-41a9-b496-bf6577afbfed-kube-api-access-hljr4\") pod \"redhat-marketplace-c44sb\" (UID: \"15b454e0-b590-41a9-b496-bf6577afbfed\") " pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.138548 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hljr4\" (UniqueName: \"kubernetes.io/projected/15b454e0-b590-41a9-b496-bf6577afbfed-kube-api-access-hljr4\") pod \"redhat-marketplace-c44sb\" (UID: \"15b454e0-b590-41a9-b496-bf6577afbfed\") " pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.138626 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15b454e0-b590-41a9-b496-bf6577afbfed-utilities\") pod \"redhat-marketplace-c44sb\" (UID: \"15b454e0-b590-41a9-b496-bf6577afbfed\") " pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.138667 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15b454e0-b590-41a9-b496-bf6577afbfed-catalog-content\") pod \"redhat-marketplace-c44sb\" (UID: \"15b454e0-b590-41a9-b496-bf6577afbfed\") " pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.139219 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/15b454e0-b590-41a9-b496-bf6577afbfed-catalog-content\") pod \"redhat-marketplace-c44sb\" (UID: \"15b454e0-b590-41a9-b496-bf6577afbfed\") " pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.139330 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/15b454e0-b590-41a9-b496-bf6577afbfed-utilities\") pod \"redhat-marketplace-c44sb\" (UID: \"15b454e0-b590-41a9-b496-bf6577afbfed\") " pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.156584 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hljr4\" (UniqueName: \"kubernetes.io/projected/15b454e0-b590-41a9-b496-bf6577afbfed-kube-api-access-hljr4\") pod \"redhat-marketplace-c44sb\" (UID: \"15b454e0-b590-41a9-b496-bf6577afbfed\") " pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.272858 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.507972 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c44sb"] Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.574045 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d555ec7-2800-45a7-a1a2-27b1a435b0d4" path="/var/lib/kubelet/pods/0d555ec7-2800-45a7-a1a2-27b1a435b0d4/volumes" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.574938 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c9ccdf7-acc0-481a-b1de-5172e16ef22c" path="/var/lib/kubelet/pods/2c9ccdf7-acc0-481a-b1de-5172e16ef22c/volumes" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.575515 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a3b7592-0bdb-4655-87a3-86ecacc6a0fa" path="/var/lib/kubelet/pods/3a3b7592-0bdb-4655-87a3-86ecacc6a0fa/volumes" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.576661 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e183218-08a1-4890-a6af-d56f3c7db2be" path="/var/lib/kubelet/pods/3e183218-08a1-4890-a6af-d56f3c7db2be/volumes" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.577220 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6da7528-9b01-40b4-88d3-e78a3c178300" path="/var/lib/kubelet/pods/a6da7528-9b01-40b4-88d3-e78a3c178300/volumes" Dec 04 10:22:22 crc kubenswrapper[4943]: I1204 10:22:22.578264 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d22ad643-0a77-49fb-a7f1-54603fe0f82d" path="/var/lib/kubelet/pods/d22ad643-0a77-49fb-a7f1-54603fe0f82d/volumes" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.135545 4943 generic.go:334] "Generic (PLEG): container finished" podID="15b454e0-b590-41a9-b496-bf6577afbfed" containerID="c1fd2902a8e602260c8c8a9a6f0b1ea3ce728a235b4971a3b867a30e8c537846" exitCode=0 Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.135648 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c44sb" event={"ID":"15b454e0-b590-41a9-b496-bf6577afbfed","Type":"ContainerDied","Data":"c1fd2902a8e602260c8c8a9a6f0b1ea3ce728a235b4971a3b867a30e8c537846"} Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.135688 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c44sb" event={"ID":"15b454e0-b590-41a9-b496-bf6577afbfed","Type":"ContainerStarted","Data":"9c7746bfe86d8fa61a500431849a511e84e647a79db5061cca3d7de993df5521"} Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.343075 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-blf86"] Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.344398 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.346115 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.353745 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-blf86"] Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.459935 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqgs4\" (UniqueName: \"kubernetes.io/projected/facfc1e3-6604-4ec2-b320-b0508d57414b-kube-api-access-kqgs4\") pod \"redhat-operators-blf86\" (UID: \"facfc1e3-6604-4ec2-b320-b0508d57414b\") " pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.460052 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/facfc1e3-6604-4ec2-b320-b0508d57414b-catalog-content\") pod \"redhat-operators-blf86\" (UID: \"facfc1e3-6604-4ec2-b320-b0508d57414b\") " pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.460092 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/facfc1e3-6604-4ec2-b320-b0508d57414b-utilities\") pod \"redhat-operators-blf86\" (UID: \"facfc1e3-6604-4ec2-b320-b0508d57414b\") " pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.561290 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqgs4\" (UniqueName: \"kubernetes.io/projected/facfc1e3-6604-4ec2-b320-b0508d57414b-kube-api-access-kqgs4\") pod \"redhat-operators-blf86\" (UID: \"facfc1e3-6604-4ec2-b320-b0508d57414b\") " pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.561365 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/facfc1e3-6604-4ec2-b320-b0508d57414b-catalog-content\") pod \"redhat-operators-blf86\" (UID: \"facfc1e3-6604-4ec2-b320-b0508d57414b\") " pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.561399 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/facfc1e3-6604-4ec2-b320-b0508d57414b-utilities\") pod \"redhat-operators-blf86\" (UID: \"facfc1e3-6604-4ec2-b320-b0508d57414b\") " pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.561930 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/facfc1e3-6604-4ec2-b320-b0508d57414b-catalog-content\") pod \"redhat-operators-blf86\" (UID: \"facfc1e3-6604-4ec2-b320-b0508d57414b\") " pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.561955 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/facfc1e3-6604-4ec2-b320-b0508d57414b-utilities\") pod \"redhat-operators-blf86\" (UID: \"facfc1e3-6604-4ec2-b320-b0508d57414b\") " pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.580754 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqgs4\" (UniqueName: \"kubernetes.io/projected/facfc1e3-6604-4ec2-b320-b0508d57414b-kube-api-access-kqgs4\") pod \"redhat-operators-blf86\" (UID: \"facfc1e3-6604-4ec2-b320-b0508d57414b\") " pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.662143 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:23 crc kubenswrapper[4943]: I1204 10:22:23.863536 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-blf86"] Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.143247 4943 generic.go:334] "Generic (PLEG): container finished" podID="facfc1e3-6604-4ec2-b320-b0508d57414b" containerID="7dce9d3648679c86c6e1c49e2b23ece4a22bd0b0fad32faac890437bb8561d5f" exitCode=0 Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.143330 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-blf86" event={"ID":"facfc1e3-6604-4ec2-b320-b0508d57414b","Type":"ContainerDied","Data":"7dce9d3648679c86c6e1c49e2b23ece4a22bd0b0fad32faac890437bb8561d5f"} Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.143541 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-blf86" event={"ID":"facfc1e3-6604-4ec2-b320-b0508d57414b","Type":"ContainerStarted","Data":"9be4ba53420e1380bd11ae21f9cc8df0dadfd762296d14839c4d0f38f4e66d62"} Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.359331 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vbmdp"] Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.360802 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.363362 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.375120 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vbmdp"] Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.472371 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46236d49-e7f4-49aa-bd3e-e770c260553a-catalog-content\") pod \"community-operators-vbmdp\" (UID: \"46236d49-e7f4-49aa-bd3e-e770c260553a\") " pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.473053 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46236d49-e7f4-49aa-bd3e-e770c260553a-utilities\") pod \"community-operators-vbmdp\" (UID: \"46236d49-e7f4-49aa-bd3e-e770c260553a\") " pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.473254 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlfsw\" (UniqueName: \"kubernetes.io/projected/46236d49-e7f4-49aa-bd3e-e770c260553a-kube-api-access-tlfsw\") pod \"community-operators-vbmdp\" (UID: \"46236d49-e7f4-49aa-bd3e-e770c260553a\") " pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.575015 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46236d49-e7f4-49aa-bd3e-e770c260553a-catalog-content\") pod \"community-operators-vbmdp\" (UID: \"46236d49-e7f4-49aa-bd3e-e770c260553a\") " pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.575697 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46236d49-e7f4-49aa-bd3e-e770c260553a-catalog-content\") pod \"community-operators-vbmdp\" (UID: \"46236d49-e7f4-49aa-bd3e-e770c260553a\") " pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.575856 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46236d49-e7f4-49aa-bd3e-e770c260553a-utilities\") pod \"community-operators-vbmdp\" (UID: \"46236d49-e7f4-49aa-bd3e-e770c260553a\") " pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.575997 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlfsw\" (UniqueName: \"kubernetes.io/projected/46236d49-e7f4-49aa-bd3e-e770c260553a-kube-api-access-tlfsw\") pod \"community-operators-vbmdp\" (UID: \"46236d49-e7f4-49aa-bd3e-e770c260553a\") " pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.576178 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46236d49-e7f4-49aa-bd3e-e770c260553a-utilities\") pod \"community-operators-vbmdp\" (UID: \"46236d49-e7f4-49aa-bd3e-e770c260553a\") " pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.594093 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlfsw\" (UniqueName: \"kubernetes.io/projected/46236d49-e7f4-49aa-bd3e-e770c260553a-kube-api-access-tlfsw\") pod \"community-operators-vbmdp\" (UID: \"46236d49-e7f4-49aa-bd3e-e770c260553a\") " pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:24 crc kubenswrapper[4943]: I1204 10:22:24.681293 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.125475 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vbmdp"] Dec 04 10:22:25 crc kubenswrapper[4943]: W1204 10:22:25.130407 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46236d49_e7f4_49aa_bd3e_e770c260553a.slice/crio-430cf088fe2666d8886f364a6b6bfcaf4a594f3ffda7f8d9afd3095bc8ebbfc5 WatchSource:0}: Error finding container 430cf088fe2666d8886f364a6b6bfcaf4a594f3ffda7f8d9afd3095bc8ebbfc5: Status 404 returned error can't find the container with id 430cf088fe2666d8886f364a6b6bfcaf4a594f3ffda7f8d9afd3095bc8ebbfc5 Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.149972 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vbmdp" event={"ID":"46236d49-e7f4-49aa-bd3e-e770c260553a","Type":"ContainerStarted","Data":"430cf088fe2666d8886f364a6b6bfcaf4a594f3ffda7f8d9afd3095bc8ebbfc5"} Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.152078 4943 generic.go:334] "Generic (PLEG): container finished" podID="15b454e0-b590-41a9-b496-bf6577afbfed" containerID="13e3b52310af67cf7d989c5cbc65d44edc39a88a5acd0e24934f574725ac2f65" exitCode=0 Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.152121 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c44sb" event={"ID":"15b454e0-b590-41a9-b496-bf6577afbfed","Type":"ContainerDied","Data":"13e3b52310af67cf7d989c5cbc65d44edc39a88a5acd0e24934f574725ac2f65"} Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.747039 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v55rk"] Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.748898 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.751717 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.759360 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v55rk"] Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.793023 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6gb8\" (UniqueName: \"kubernetes.io/projected/f898614f-17cf-4133-9e79-08eff3ee3979-kube-api-access-j6gb8\") pod \"certified-operators-v55rk\" (UID: \"f898614f-17cf-4133-9e79-08eff3ee3979\") " pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.793070 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f898614f-17cf-4133-9e79-08eff3ee3979-catalog-content\") pod \"certified-operators-v55rk\" (UID: \"f898614f-17cf-4133-9e79-08eff3ee3979\") " pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.793153 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f898614f-17cf-4133-9e79-08eff3ee3979-utilities\") pod \"certified-operators-v55rk\" (UID: \"f898614f-17cf-4133-9e79-08eff3ee3979\") " pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.894872 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f898614f-17cf-4133-9e79-08eff3ee3979-catalog-content\") pod \"certified-operators-v55rk\" (UID: \"f898614f-17cf-4133-9e79-08eff3ee3979\") " pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.894940 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f898614f-17cf-4133-9e79-08eff3ee3979-utilities\") pod \"certified-operators-v55rk\" (UID: \"f898614f-17cf-4133-9e79-08eff3ee3979\") " pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.895010 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6gb8\" (UniqueName: \"kubernetes.io/projected/f898614f-17cf-4133-9e79-08eff3ee3979-kube-api-access-j6gb8\") pod \"certified-operators-v55rk\" (UID: \"f898614f-17cf-4133-9e79-08eff3ee3979\") " pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.895476 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f898614f-17cf-4133-9e79-08eff3ee3979-catalog-content\") pod \"certified-operators-v55rk\" (UID: \"f898614f-17cf-4133-9e79-08eff3ee3979\") " pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.895521 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f898614f-17cf-4133-9e79-08eff3ee3979-utilities\") pod \"certified-operators-v55rk\" (UID: \"f898614f-17cf-4133-9e79-08eff3ee3979\") " pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:25 crc kubenswrapper[4943]: I1204 10:22:25.914809 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6gb8\" (UniqueName: \"kubernetes.io/projected/f898614f-17cf-4133-9e79-08eff3ee3979-kube-api-access-j6gb8\") pod \"certified-operators-v55rk\" (UID: \"f898614f-17cf-4133-9e79-08eff3ee3979\") " pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:26 crc kubenswrapper[4943]: I1204 10:22:26.073915 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:26 crc kubenswrapper[4943]: I1204 10:22:26.178591 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-blf86" event={"ID":"facfc1e3-6604-4ec2-b320-b0508d57414b","Type":"ContainerStarted","Data":"7873b6ca4538edd0d1c5a4f0bd9ad73b89ee31700b71da16ed83e4799e65be66"} Dec 04 10:22:26 crc kubenswrapper[4943]: I1204 10:22:26.233839 4943 generic.go:334] "Generic (PLEG): container finished" podID="46236d49-e7f4-49aa-bd3e-e770c260553a" containerID="03141d3849791a979d6769b2e211243a7025fed6aa20bdd5bb589f8dd0850d60" exitCode=0 Dec 04 10:22:26 crc kubenswrapper[4943]: I1204 10:22:26.233886 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vbmdp" event={"ID":"46236d49-e7f4-49aa-bd3e-e770c260553a","Type":"ContainerDied","Data":"03141d3849791a979d6769b2e211243a7025fed6aa20bdd5bb589f8dd0850d60"} Dec 04 10:22:26 crc kubenswrapper[4943]: I1204 10:22:26.539012 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v55rk"] Dec 04 10:22:26 crc kubenswrapper[4943]: W1204 10:22:26.558318 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf898614f_17cf_4133_9e79_08eff3ee3979.slice/crio-047826af6a1ae940ca41075fb344821b6edc45b561d72571a0027052b4f5194b WatchSource:0}: Error finding container 047826af6a1ae940ca41075fb344821b6edc45b561d72571a0027052b4f5194b: Status 404 returned error can't find the container with id 047826af6a1ae940ca41075fb344821b6edc45b561d72571a0027052b4f5194b Dec 04 10:22:27 crc kubenswrapper[4943]: I1204 10:22:27.241732 4943 generic.go:334] "Generic (PLEG): container finished" podID="facfc1e3-6604-4ec2-b320-b0508d57414b" containerID="7873b6ca4538edd0d1c5a4f0bd9ad73b89ee31700b71da16ed83e4799e65be66" exitCode=0 Dec 04 10:22:27 crc kubenswrapper[4943]: I1204 10:22:27.241834 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-blf86" event={"ID":"facfc1e3-6604-4ec2-b320-b0508d57414b","Type":"ContainerDied","Data":"7873b6ca4538edd0d1c5a4f0bd9ad73b89ee31700b71da16ed83e4799e65be66"} Dec 04 10:22:27 crc kubenswrapper[4943]: I1204 10:22:27.247122 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c44sb" event={"ID":"15b454e0-b590-41a9-b496-bf6577afbfed","Type":"ContainerStarted","Data":"30cefb4651689decb5825d71798f0326c49fe3367bcd32ef86523e1f09989488"} Dec 04 10:22:27 crc kubenswrapper[4943]: I1204 10:22:27.248103 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v55rk" event={"ID":"f898614f-17cf-4133-9e79-08eff3ee3979","Type":"ContainerStarted","Data":"047826af6a1ae940ca41075fb344821b6edc45b561d72571a0027052b4f5194b"} Dec 04 10:22:27 crc kubenswrapper[4943]: I1204 10:22:27.285789 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c44sb" podStartSLOduration=3.380556021 podStartE2EDuration="6.285769934s" podCreationTimestamp="2025-12-04 10:22:21 +0000 UTC" firstStartedPulling="2025-12-04 10:22:23.137716227 +0000 UTC m=+391.726692095" lastFinishedPulling="2025-12-04 10:22:26.04293015 +0000 UTC m=+394.631906008" observedRunningTime="2025-12-04 10:22:27.282477638 +0000 UTC m=+395.871453506" watchObservedRunningTime="2025-12-04 10:22:27.285769934 +0000 UTC m=+395.874745792" Dec 04 10:22:28 crc kubenswrapper[4943]: I1204 10:22:28.254055 4943 generic.go:334] "Generic (PLEG): container finished" podID="f898614f-17cf-4133-9e79-08eff3ee3979" containerID="dfc0fd6f87cdde54206a1c1521c69380974337698ab0ea540b94874d5cf84d30" exitCode=0 Dec 04 10:22:28 crc kubenswrapper[4943]: I1204 10:22:28.255919 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v55rk" event={"ID":"f898614f-17cf-4133-9e79-08eff3ee3979","Type":"ContainerDied","Data":"dfc0fd6f87cdde54206a1c1521c69380974337698ab0ea540b94874d5cf84d30"} Dec 04 10:22:28 crc kubenswrapper[4943]: I1204 10:22:28.594748 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-76b976544-pfp2x"] Dec 04 10:22:28 crc kubenswrapper[4943]: I1204 10:22:28.594956 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" podUID="a87d9564-2875-4550-af91-a9257d9ae83c" containerName="controller-manager" containerID="cri-o://52a18460e2920491a24acab6877cca30f8730bf4a2260abe4aa1754e8cbffde6" gracePeriod=30 Dec 04 10:22:28 crc kubenswrapper[4943]: I1204 10:22:28.602648 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx"] Dec 04 10:22:28 crc kubenswrapper[4943]: I1204 10:22:28.602891 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" podUID="1922c24e-5702-4d9a-a5b9-2abaa98871f4" containerName="route-controller-manager" containerID="cri-o://a1145e8481ac2d807a25e1e8948571030bd5398f4b6e7d7ea4d5018e7f83cdaf" gracePeriod=30 Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.274436 4943 generic.go:334] "Generic (PLEG): container finished" podID="1922c24e-5702-4d9a-a5b9-2abaa98871f4" containerID="a1145e8481ac2d807a25e1e8948571030bd5398f4b6e7d7ea4d5018e7f83cdaf" exitCode=0 Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.274907 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" event={"ID":"1922c24e-5702-4d9a-a5b9-2abaa98871f4","Type":"ContainerDied","Data":"a1145e8481ac2d807a25e1e8948571030bd5398f4b6e7d7ea4d5018e7f83cdaf"} Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.277539 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vbmdp" event={"ID":"46236d49-e7f4-49aa-bd3e-e770c260553a","Type":"ContainerStarted","Data":"f25f0b50f03ff6dbb469ebddd75dd2bc1f168f0ad8962be9fa1245100badc1d7"} Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.279937 4943 generic.go:334] "Generic (PLEG): container finished" podID="a87d9564-2875-4550-af91-a9257d9ae83c" containerID="52a18460e2920491a24acab6877cca30f8730bf4a2260abe4aa1754e8cbffde6" exitCode=0 Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.280021 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" event={"ID":"a87d9564-2875-4550-af91-a9257d9ae83c","Type":"ContainerDied","Data":"52a18460e2920491a24acab6877cca30f8730bf4a2260abe4aa1754e8cbffde6"} Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.291230 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-blf86" event={"ID":"facfc1e3-6604-4ec2-b320-b0508d57414b","Type":"ContainerStarted","Data":"5de7ed5ecdf71890c32ea9c4ae5f249701afd67974bdcafb539cb2b8e1a418ed"} Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.609290 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.614113 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.639257 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-blf86" podStartSLOduration=3.044481625 podStartE2EDuration="6.639228118s" podCreationTimestamp="2025-12-04 10:22:23 +0000 UTC" firstStartedPulling="2025-12-04 10:22:24.149265442 +0000 UTC m=+392.738241330" lastFinishedPulling="2025-12-04 10:22:27.744011955 +0000 UTC m=+396.332987823" observedRunningTime="2025-12-04 10:22:29.324809202 +0000 UTC m=+397.913785060" watchObservedRunningTime="2025-12-04 10:22:29.639228118 +0000 UTC m=+398.228203986" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.732742 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-config\") pod \"a87d9564-2875-4550-af91-a9257d9ae83c\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.732826 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1922c24e-5702-4d9a-a5b9-2abaa98871f4-serving-cert\") pod \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.732861 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-proxy-ca-bundles\") pod \"a87d9564-2875-4550-af91-a9257d9ae83c\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.732883 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1922c24e-5702-4d9a-a5b9-2abaa98871f4-client-ca\") pod \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.732919 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-client-ca\") pod \"a87d9564-2875-4550-af91-a9257d9ae83c\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.732966 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clhnw\" (UniqueName: \"kubernetes.io/projected/1922c24e-5702-4d9a-a5b9-2abaa98871f4-kube-api-access-clhnw\") pod \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.732998 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1922c24e-5702-4d9a-a5b9-2abaa98871f4-config\") pod \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\" (UID: \"1922c24e-5702-4d9a-a5b9-2abaa98871f4\") " Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.733024 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdwb6\" (UniqueName: \"kubernetes.io/projected/a87d9564-2875-4550-af91-a9257d9ae83c-kube-api-access-wdwb6\") pod \"a87d9564-2875-4550-af91-a9257d9ae83c\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.733071 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87d9564-2875-4550-af91-a9257d9ae83c-serving-cert\") pod \"a87d9564-2875-4550-af91-a9257d9ae83c\" (UID: \"a87d9564-2875-4550-af91-a9257d9ae83c\") " Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.734384 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1922c24e-5702-4d9a-a5b9-2abaa98871f4-client-ca" (OuterVolumeSpecName: "client-ca") pod "1922c24e-5702-4d9a-a5b9-2abaa98871f4" (UID: "1922c24e-5702-4d9a-a5b9-2abaa98871f4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.734975 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-config" (OuterVolumeSpecName: "config") pod "a87d9564-2875-4550-af91-a9257d9ae83c" (UID: "a87d9564-2875-4550-af91-a9257d9ae83c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.735022 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1922c24e-5702-4d9a-a5b9-2abaa98871f4-config" (OuterVolumeSpecName: "config") pod "1922c24e-5702-4d9a-a5b9-2abaa98871f4" (UID: "1922c24e-5702-4d9a-a5b9-2abaa98871f4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.736571 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a87d9564-2875-4550-af91-a9257d9ae83c" (UID: "a87d9564-2875-4550-af91-a9257d9ae83c"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.736966 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-client-ca" (OuterVolumeSpecName: "client-ca") pod "a87d9564-2875-4550-af91-a9257d9ae83c" (UID: "a87d9564-2875-4550-af91-a9257d9ae83c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.739868 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87d9564-2875-4550-af91-a9257d9ae83c-kube-api-access-wdwb6" (OuterVolumeSpecName: "kube-api-access-wdwb6") pod "a87d9564-2875-4550-af91-a9257d9ae83c" (UID: "a87d9564-2875-4550-af91-a9257d9ae83c"). InnerVolumeSpecName "kube-api-access-wdwb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.740597 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1922c24e-5702-4d9a-a5b9-2abaa98871f4-kube-api-access-clhnw" (OuterVolumeSpecName: "kube-api-access-clhnw") pod "1922c24e-5702-4d9a-a5b9-2abaa98871f4" (UID: "1922c24e-5702-4d9a-a5b9-2abaa98871f4"). InnerVolumeSpecName "kube-api-access-clhnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.740679 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87d9564-2875-4550-af91-a9257d9ae83c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a87d9564-2875-4550-af91-a9257d9ae83c" (UID: "a87d9564-2875-4550-af91-a9257d9ae83c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.744322 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1922c24e-5702-4d9a-a5b9-2abaa98871f4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1922c24e-5702-4d9a-a5b9-2abaa98871f4" (UID: "1922c24e-5702-4d9a-a5b9-2abaa98871f4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.834494 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1922c24e-5702-4d9a-a5b9-2abaa98871f4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.834527 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1922c24e-5702-4d9a-a5b9-2abaa98871f4-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.834539 4943 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.834551 4943 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.834565 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clhnw\" (UniqueName: \"kubernetes.io/projected/1922c24e-5702-4d9a-a5b9-2abaa98871f4-kube-api-access-clhnw\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.834576 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1922c24e-5702-4d9a-a5b9-2abaa98871f4-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.834587 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdwb6\" (UniqueName: \"kubernetes.io/projected/a87d9564-2875-4550-af91-a9257d9ae83c-kube-api-access-wdwb6\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.834596 4943 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a87d9564-2875-4550-af91-a9257d9ae83c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:29 crc kubenswrapper[4943]: I1204 10:22:29.834605 4943 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a87d9564-2875-4550-af91-a9257d9ae83c-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.298424 4943 generic.go:334] "Generic (PLEG): container finished" podID="46236d49-e7f4-49aa-bd3e-e770c260553a" containerID="f25f0b50f03ff6dbb469ebddd75dd2bc1f168f0ad8962be9fa1245100badc1d7" exitCode=0 Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.298504 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vbmdp" event={"ID":"46236d49-e7f4-49aa-bd3e-e770c260553a","Type":"ContainerDied","Data":"f25f0b50f03ff6dbb469ebddd75dd2bc1f168f0ad8962be9fa1245100badc1d7"} Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.306656 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" event={"ID":"a87d9564-2875-4550-af91-a9257d9ae83c","Type":"ContainerDied","Data":"705755991edfca6c90c14731e15929cbc113ee14eaca05f42170171fafb9d3f8"} Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.306746 4943 scope.go:117] "RemoveContainer" containerID="52a18460e2920491a24acab6877cca30f8730bf4a2260abe4aa1754e8cbffde6" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.306863 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76b976544-pfp2x" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.309379 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" event={"ID":"1922c24e-5702-4d9a-a5b9-2abaa98871f4","Type":"ContainerDied","Data":"c49aff73f6d86281a4db981c1529245153182284e9951e9f222dfadbf36d39d4"} Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.309399 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.350814 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp"] Dec 04 10:22:30 crc kubenswrapper[4943]: E1204 10:22:30.351273 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87d9564-2875-4550-af91-a9257d9ae83c" containerName="controller-manager" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.351388 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87d9564-2875-4550-af91-a9257d9ae83c" containerName="controller-manager" Dec 04 10:22:30 crc kubenswrapper[4943]: E1204 10:22:30.351473 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1922c24e-5702-4d9a-a5b9-2abaa98871f4" containerName="route-controller-manager" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.351552 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1922c24e-5702-4d9a-a5b9-2abaa98871f4" containerName="route-controller-manager" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.351765 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1922c24e-5702-4d9a-a5b9-2abaa98871f4" containerName="route-controller-manager" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.351860 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a87d9564-2875-4550-af91-a9257d9ae83c" containerName="controller-manager" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.353113 4943 scope.go:117] "RemoveContainer" containerID="a1145e8481ac2d807a25e1e8948571030bd5398f4b6e7d7ea4d5018e7f83cdaf" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.354528 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.357215 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.357396 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.357481 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-654c89ff58-sk2bz"] Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.358353 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.361285 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.362825 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.363149 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.364261 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.364643 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.371569 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.371762 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.371893 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.372903 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx"] Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.373073 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.383772 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8d8c54746-hkqbx"] Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.384165 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.397791 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-654c89ff58-sk2bz"] Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.433093 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp"] Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.435807 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.440405 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-76b976544-pfp2x"] Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.445114 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-76b976544-pfp2x"] Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.445959 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b8bf6b0-7046-4216-83db-7db180db8659-config\") pod \"route-controller-manager-fff48484f-xw8cp\" (UID: \"9b8bf6b0-7046-4216-83db-7db180db8659\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.446008 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea2f7791-96df-419b-90a6-9128281edf47-config\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.446039 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea2f7791-96df-419b-90a6-9128281edf47-serving-cert\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.446090 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ea2f7791-96df-419b-90a6-9128281edf47-proxy-ca-bundles\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.446117 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b8bf6b0-7046-4216-83db-7db180db8659-client-ca\") pod \"route-controller-manager-fff48484f-xw8cp\" (UID: \"9b8bf6b0-7046-4216-83db-7db180db8659\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.446147 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b8bf6b0-7046-4216-83db-7db180db8659-serving-cert\") pod \"route-controller-manager-fff48484f-xw8cp\" (UID: \"9b8bf6b0-7046-4216-83db-7db180db8659\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.446417 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4r6z\" (UniqueName: \"kubernetes.io/projected/9b8bf6b0-7046-4216-83db-7db180db8659-kube-api-access-j4r6z\") pod \"route-controller-manager-fff48484f-xw8cp\" (UID: \"9b8bf6b0-7046-4216-83db-7db180db8659\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.446455 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea2f7791-96df-419b-90a6-9128281edf47-client-ca\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.446477 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nbvq\" (UniqueName: \"kubernetes.io/projected/ea2f7791-96df-419b-90a6-9128281edf47-kube-api-access-6nbvq\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.548262 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b8bf6b0-7046-4216-83db-7db180db8659-serving-cert\") pod \"route-controller-manager-fff48484f-xw8cp\" (UID: \"9b8bf6b0-7046-4216-83db-7db180db8659\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.548358 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4r6z\" (UniqueName: \"kubernetes.io/projected/9b8bf6b0-7046-4216-83db-7db180db8659-kube-api-access-j4r6z\") pod \"route-controller-manager-fff48484f-xw8cp\" (UID: \"9b8bf6b0-7046-4216-83db-7db180db8659\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.548409 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea2f7791-96df-419b-90a6-9128281edf47-client-ca\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.548432 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nbvq\" (UniqueName: \"kubernetes.io/projected/ea2f7791-96df-419b-90a6-9128281edf47-kube-api-access-6nbvq\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.548456 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b8bf6b0-7046-4216-83db-7db180db8659-config\") pod \"route-controller-manager-fff48484f-xw8cp\" (UID: \"9b8bf6b0-7046-4216-83db-7db180db8659\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.548477 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea2f7791-96df-419b-90a6-9128281edf47-config\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.548504 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea2f7791-96df-419b-90a6-9128281edf47-serving-cert\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.548555 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ea2f7791-96df-419b-90a6-9128281edf47-proxy-ca-bundles\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.548573 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b8bf6b0-7046-4216-83db-7db180db8659-client-ca\") pod \"route-controller-manager-fff48484f-xw8cp\" (UID: \"9b8bf6b0-7046-4216-83db-7db180db8659\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.549736 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b8bf6b0-7046-4216-83db-7db180db8659-client-ca\") pod \"route-controller-manager-fff48484f-xw8cp\" (UID: \"9b8bf6b0-7046-4216-83db-7db180db8659\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.550272 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ea2f7791-96df-419b-90a6-9128281edf47-client-ca\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.550354 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b8bf6b0-7046-4216-83db-7db180db8659-config\") pod \"route-controller-manager-fff48484f-xw8cp\" (UID: \"9b8bf6b0-7046-4216-83db-7db180db8659\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.550551 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ea2f7791-96df-419b-90a6-9128281edf47-proxy-ca-bundles\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.551468 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea2f7791-96df-419b-90a6-9128281edf47-config\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.553581 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b8bf6b0-7046-4216-83db-7db180db8659-serving-cert\") pod \"route-controller-manager-fff48484f-xw8cp\" (UID: \"9b8bf6b0-7046-4216-83db-7db180db8659\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.569033 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ea2f7791-96df-419b-90a6-9128281edf47-serving-cert\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.570063 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1922c24e-5702-4d9a-a5b9-2abaa98871f4" path="/var/lib/kubelet/pods/1922c24e-5702-4d9a-a5b9-2abaa98871f4/volumes" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.570700 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a87d9564-2875-4550-af91-a9257d9ae83c" path="/var/lib/kubelet/pods/a87d9564-2875-4550-af91-a9257d9ae83c/volumes" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.571468 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4r6z\" (UniqueName: \"kubernetes.io/projected/9b8bf6b0-7046-4216-83db-7db180db8659-kube-api-access-j4r6z\") pod \"route-controller-manager-fff48484f-xw8cp\" (UID: \"9b8bf6b0-7046-4216-83db-7db180db8659\") " pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.572411 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nbvq\" (UniqueName: \"kubernetes.io/projected/ea2f7791-96df-419b-90a6-9128281edf47-kube-api-access-6nbvq\") pod \"controller-manager-654c89ff58-sk2bz\" (UID: \"ea2f7791-96df-419b-90a6-9128281edf47\") " pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.680137 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.702931 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.868987 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp"] Dec 04 10:22:30 crc kubenswrapper[4943]: I1204 10:22:30.941033 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-654c89ff58-sk2bz"] Dec 04 10:22:30 crc kubenswrapper[4943]: W1204 10:22:30.952044 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea2f7791_96df_419b_90a6_9128281edf47.slice/crio-0bfa8831f005daaea6f6ca73ffa190d220fa3e9752d725cec87d8be4a24a3b8a WatchSource:0}: Error finding container 0bfa8831f005daaea6f6ca73ffa190d220fa3e9752d725cec87d8be4a24a3b8a: Status 404 returned error can't find the container with id 0bfa8831f005daaea6f6ca73ffa190d220fa3e9752d725cec87d8be4a24a3b8a Dec 04 10:22:31 crc kubenswrapper[4943]: I1204 10:22:31.318370 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" event={"ID":"9b8bf6b0-7046-4216-83db-7db180db8659","Type":"ContainerStarted","Data":"f93ec8ba66d64d21a5331705d881438b06eb57cab2620875d2ae79f72c4766eb"} Dec 04 10:22:31 crc kubenswrapper[4943]: I1204 10:22:31.320029 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" event={"ID":"ea2f7791-96df-419b-90a6-9128281edf47","Type":"ContainerStarted","Data":"0bfa8831f005daaea6f6ca73ffa190d220fa3e9752d725cec87d8be4a24a3b8a"} Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.273678 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.273996 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.328742 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vbmdp" event={"ID":"46236d49-e7f4-49aa-bd3e-e770c260553a","Type":"ContainerStarted","Data":"8295b4d3a9c35b051e62fc7eb6208f610c0e700a87031d61159276bf4f887387"} Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.331060 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.333826 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" event={"ID":"ea2f7791-96df-419b-90a6-9128281edf47","Type":"ContainerStarted","Data":"10cac71fe28078fca603eaa00af4408600ae83a0877eff04e1393972b7d6b8fd"} Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.334051 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.336221 4943 generic.go:334] "Generic (PLEG): container finished" podID="f898614f-17cf-4133-9e79-08eff3ee3979" containerID="a5d8c83a060a1fa01d5d926a5a6d5c699fcb815e0b742ce88c15020121ebcd2d" exitCode=0 Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.336339 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v55rk" event={"ID":"f898614f-17cf-4133-9e79-08eff3ee3979","Type":"ContainerDied","Data":"a5d8c83a060a1fa01d5d926a5a6d5c699fcb815e0b742ce88c15020121ebcd2d"} Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.338583 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" event={"ID":"9b8bf6b0-7046-4216-83db-7db180db8659","Type":"ContainerStarted","Data":"1a96f2d67339c9e30efbee7eadae6d9f4c9d7a9870d2429716d4afc2550cb7a7"} Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.339525 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.342292 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.346992 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.356359 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vbmdp" podStartSLOduration=3.268086274 podStartE2EDuration="8.356340038s" podCreationTimestamp="2025-12-04 10:22:24 +0000 UTC" firstStartedPulling="2025-12-04 10:22:26.241420513 +0000 UTC m=+394.830396381" lastFinishedPulling="2025-12-04 10:22:31.329674277 +0000 UTC m=+399.918650145" observedRunningTime="2025-12-04 10:22:32.351702024 +0000 UTC m=+400.940677892" watchObservedRunningTime="2025-12-04 10:22:32.356340038 +0000 UTC m=+400.945315906" Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.374293 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-654c89ff58-sk2bz" podStartSLOduration=4.374259234 podStartE2EDuration="4.374259234s" podCreationTimestamp="2025-12-04 10:22:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:22:32.369045874 +0000 UTC m=+400.958021752" watchObservedRunningTime="2025-12-04 10:22:32.374259234 +0000 UTC m=+400.963235112" Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.389839 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c44sb" Dec 04 10:22:32 crc kubenswrapper[4943]: I1204 10:22:32.473751 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-fff48484f-xw8cp" podStartSLOduration=4.4737274320000004 podStartE2EDuration="4.473727432s" podCreationTimestamp="2025-12-04 10:22:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:22:32.453496889 +0000 UTC m=+401.042472767" watchObservedRunningTime="2025-12-04 10:22:32.473727432 +0000 UTC m=+401.062703300" Dec 04 10:22:33 crc kubenswrapper[4943]: I1204 10:22:33.364800 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v55rk" event={"ID":"f898614f-17cf-4133-9e79-08eff3ee3979","Type":"ContainerStarted","Data":"d82dbfad0173069a79d4965066895ecc8b00d14a8ffdebfc7a1f644e01497946"} Dec 04 10:22:33 crc kubenswrapper[4943]: I1204 10:22:33.395522 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v55rk" podStartSLOduration=3.7978375 podStartE2EDuration="8.395500009s" podCreationTimestamp="2025-12-04 10:22:25 +0000 UTC" firstStartedPulling="2025-12-04 10:22:28.25783709 +0000 UTC m=+396.846812968" lastFinishedPulling="2025-12-04 10:22:32.855499609 +0000 UTC m=+401.444475477" observedRunningTime="2025-12-04 10:22:33.385706626 +0000 UTC m=+401.974682494" watchObservedRunningTime="2025-12-04 10:22:33.395500009 +0000 UTC m=+401.984475877" Dec 04 10:22:33 crc kubenswrapper[4943]: I1204 10:22:33.663464 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:33 crc kubenswrapper[4943]: I1204 10:22:33.663671 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:34 crc kubenswrapper[4943]: I1204 10:22:34.682852 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:34 crc kubenswrapper[4943]: I1204 10:22:34.683422 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:34 crc kubenswrapper[4943]: I1204 10:22:34.705956 4943 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-blf86" podUID="facfc1e3-6604-4ec2-b320-b0508d57414b" containerName="registry-server" probeResult="failure" output=< Dec 04 10:22:34 crc kubenswrapper[4943]: timeout: failed to connect service ":50051" within 1s Dec 04 10:22:34 crc kubenswrapper[4943]: > Dec 04 10:22:34 crc kubenswrapper[4943]: I1204 10:22:34.726384 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:36 crc kubenswrapper[4943]: I1204 10:22:36.075135 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:36 crc kubenswrapper[4943]: I1204 10:22:36.075862 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:36 crc kubenswrapper[4943]: I1204 10:22:36.120906 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:37 crc kubenswrapper[4943]: I1204 10:22:37.433474 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v55rk" Dec 04 10:22:43 crc kubenswrapper[4943]: I1204 10:22:43.704026 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:43 crc kubenswrapper[4943]: I1204 10:22:43.747878 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-blf86" Dec 04 10:22:44 crc kubenswrapper[4943]: I1204 10:22:44.721386 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vbmdp" Dec 04 10:22:48 crc kubenswrapper[4943]: I1204 10:22:48.629448 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:22:48 crc kubenswrapper[4943]: I1204 10:22:48.629525 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:22:48 crc kubenswrapper[4943]: I1204 10:22:48.629583 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:22:48 crc kubenswrapper[4943]: I1204 10:22:48.630355 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"98a02a23954dbf78c645fbdf023f59c0401bc3eb6dc01df9396ba4979fbe8194"} pod="openshift-machine-config-operator/machine-config-daemon-kswzd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 10:22:48 crc kubenswrapper[4943]: I1204 10:22:48.630700 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" containerID="cri-o://98a02a23954dbf78c645fbdf023f59c0401bc3eb6dc01df9396ba4979fbe8194" gracePeriod=600 Dec 04 10:22:50 crc kubenswrapper[4943]: I1204 10:22:50.464615 4943 generic.go:334] "Generic (PLEG): container finished" podID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerID="98a02a23954dbf78c645fbdf023f59c0401bc3eb6dc01df9396ba4979fbe8194" exitCode=0 Dec 04 10:22:50 crc kubenswrapper[4943]: I1204 10:22:50.464667 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerDied","Data":"98a02a23954dbf78c645fbdf023f59c0401bc3eb6dc01df9396ba4979fbe8194"} Dec 04 10:22:50 crc kubenswrapper[4943]: I1204 10:22:50.464735 4943 scope.go:117] "RemoveContainer" containerID="9f26bd53ed38b87b818909e4bf97369bf7728f83d9aa2d93dd8fddfd4d3c5b58" Dec 04 10:22:51 crc kubenswrapper[4943]: I1204 10:22:51.473481 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerStarted","Data":"380dbb79af465e067f765975f99ff0aee03e2a47f647161e34ca81e35d30b54e"} Dec 04 10:25:18 crc kubenswrapper[4943]: I1204 10:25:18.629582 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:25:18 crc kubenswrapper[4943]: I1204 10:25:18.630307 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:25:48 crc kubenswrapper[4943]: I1204 10:25:48.628999 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:25:48 crc kubenswrapper[4943]: I1204 10:25:48.629752 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:26:18 crc kubenswrapper[4943]: I1204 10:26:18.629814 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:26:18 crc kubenswrapper[4943]: I1204 10:26:18.630412 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:26:18 crc kubenswrapper[4943]: I1204 10:26:18.630495 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:26:18 crc kubenswrapper[4943]: I1204 10:26:18.632265 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"380dbb79af465e067f765975f99ff0aee03e2a47f647161e34ca81e35d30b54e"} pod="openshift-machine-config-operator/machine-config-daemon-kswzd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 10:26:18 crc kubenswrapper[4943]: I1204 10:26:18.632377 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" containerID="cri-o://380dbb79af465e067f765975f99ff0aee03e2a47f647161e34ca81e35d30b54e" gracePeriod=600 Dec 04 10:26:19 crc kubenswrapper[4943]: I1204 10:26:19.677524 4943 generic.go:334] "Generic (PLEG): container finished" podID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerID="380dbb79af465e067f765975f99ff0aee03e2a47f647161e34ca81e35d30b54e" exitCode=0 Dec 04 10:26:19 crc kubenswrapper[4943]: I1204 10:26:19.677607 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerDied","Data":"380dbb79af465e067f765975f99ff0aee03e2a47f647161e34ca81e35d30b54e"} Dec 04 10:26:19 crc kubenswrapper[4943]: I1204 10:26:19.678341 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerStarted","Data":"2a882281b91801c61cd5c5ba5bc934cd8e35cb87485c0c878f8c0bf55706b5c5"} Dec 04 10:26:19 crc kubenswrapper[4943]: I1204 10:26:19.678406 4943 scope.go:117] "RemoveContainer" containerID="98a02a23954dbf78c645fbdf023f59c0401bc3eb6dc01df9396ba4979fbe8194" Dec 04 10:28:18 crc kubenswrapper[4943]: I1204 10:28:18.629221 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:28:18 crc kubenswrapper[4943]: I1204 10:28:18.629811 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:28:35 crc kubenswrapper[4943]: I1204 10:28:35.035517 4943 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 10:28:48 crc kubenswrapper[4943]: I1204 10:28:48.629062 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:28:48 crc kubenswrapper[4943]: I1204 10:28:48.629954 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:29:18 crc kubenswrapper[4943]: I1204 10:29:18.629417 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:29:18 crc kubenswrapper[4943]: I1204 10:29:18.630907 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:29:18 crc kubenswrapper[4943]: I1204 10:29:18.631144 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:29:18 crc kubenswrapper[4943]: I1204 10:29:18.631765 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a882281b91801c61cd5c5ba5bc934cd8e35cb87485c0c878f8c0bf55706b5c5"} pod="openshift-machine-config-operator/machine-config-daemon-kswzd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 10:29:18 crc kubenswrapper[4943]: I1204 10:29:18.631920 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" containerID="cri-o://2a882281b91801c61cd5c5ba5bc934cd8e35cb87485c0c878f8c0bf55706b5c5" gracePeriod=600 Dec 04 10:29:19 crc kubenswrapper[4943]: I1204 10:29:19.668946 4943 generic.go:334] "Generic (PLEG): container finished" podID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerID="2a882281b91801c61cd5c5ba5bc934cd8e35cb87485c0c878f8c0bf55706b5c5" exitCode=0 Dec 04 10:29:19 crc kubenswrapper[4943]: I1204 10:29:19.669003 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerDied","Data":"2a882281b91801c61cd5c5ba5bc934cd8e35cb87485c0c878f8c0bf55706b5c5"} Dec 04 10:29:19 crc kubenswrapper[4943]: I1204 10:29:19.669441 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerStarted","Data":"41363b73659bf0d33a8ed7f51e273d64bed37bf7e075e5c6bc173eb617215207"} Dec 04 10:29:19 crc kubenswrapper[4943]: I1204 10:29:19.669489 4943 scope.go:117] "RemoveContainer" containerID="380dbb79af465e067f765975f99ff0aee03e2a47f647161e34ca81e35d30b54e" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.612168 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7g4wf"] Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.613693 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="nbdb" containerID="cri-o://7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98" gracePeriod=30 Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.614079 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="sbdb" containerID="cri-o://1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284" gracePeriod=30 Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.613645 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovn-controller" containerID="cri-o://aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326" gracePeriod=30 Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.614290 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="northd" containerID="cri-o://86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744" gracePeriod=30 Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.614404 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba" gracePeriod=30 Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.614483 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="kube-rbac-proxy-node" containerID="cri-o://ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761" gracePeriod=30 Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.614557 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovn-acl-logging" containerID="cri-o://8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5" gracePeriod=30 Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.656969 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" containerID="cri-o://68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce" gracePeriod=30 Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.725214 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wkbf5_b9cb494f-7d4c-45f2-8b9b-e35c42b41c79/kube-multus/2.log" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.725908 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wkbf5_b9cb494f-7d4c-45f2-8b9b-e35c42b41c79/kube-multus/1.log" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.725957 4943 generic.go:334] "Generic (PLEG): container finished" podID="b9cb494f-7d4c-45f2-8b9b-e35c42b41c79" containerID="f48e036e9d741d48766b670629348a5e5592c0f1ef59ed8f5059e0df75057a5a" exitCode=2 Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.725992 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wkbf5" event={"ID":"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79","Type":"ContainerDied","Data":"f48e036e9d741d48766b670629348a5e5592c0f1ef59ed8f5059e0df75057a5a"} Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.726038 4943 scope.go:117] "RemoveContainer" containerID="635d7d730273976acb953de2147ab4353647915f5f9898a96c04558ab8262a1d" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.726601 4943 scope.go:117] "RemoveContainer" containerID="f48e036e9d741d48766b670629348a5e5592c0f1ef59ed8f5059e0df75057a5a" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.932392 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/3.log" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.934858 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovn-acl-logging/0.log" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.935416 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovn-controller/0.log" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.935909 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.991253 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lptpz"] Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.991800 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.991914 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.991985 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="nbdb" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.992055 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="nbdb" Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.993273 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="sbdb" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.993379 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="sbdb" Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.993456 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovn-acl-logging" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.993528 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovn-acl-logging" Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.993620 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="northd" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.993689 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="northd" Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.993955 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.994021 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.994093 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="kube-rbac-proxy-node" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.994158 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="kube-rbac-proxy-node" Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.994244 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.994332 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.994436 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovn-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.994502 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovn-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.994577 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.994643 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.994712 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="kubecfg-setup" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.994776 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="kubecfg-setup" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.995030 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.995805 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="kube-rbac-proxy-node" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.995915 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.996013 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="nbdb" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.996091 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovn-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.996162 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovn-acl-logging" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.996270 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.996336 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="northd" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.996403 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.996476 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.996539 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="sbdb" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.996600 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.996784 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.996851 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: E1204 10:29:27.996916 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:27 crc kubenswrapper[4943]: I1204 10:29:27.996986 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerName="ovnkube-controller" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.002338 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120135 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-var-lib-openvswitch\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120182 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-log-socket\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120251 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-cni-bin\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120275 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120331 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120377 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120360 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-log-socket" (OuterVolumeSpecName: "log-socket") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120403 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-openvswitch\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120420 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120444 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120489 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wh9m\" (UniqueName: \"kubernetes.io/projected/d0498f58-c99b-4fa8-a404-f7f43199c29f-kube-api-access-7wh9m\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120538 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-run-netns\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120562 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-env-overrides\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120601 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-ovn\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120632 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-kubelet\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120656 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovnkube-script-lib\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120685 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120695 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-run-ovn-kubernetes\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120720 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-cni-netd\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120740 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-systemd\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120767 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovnkube-config\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120790 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-systemd-units\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120781 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120814 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-slash\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120855 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-etc-openvswitch\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120879 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-node-log\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120912 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovn-node-metrics-cert\") pod \"d0498f58-c99b-4fa8-a404-f7f43199c29f\" (UID: \"d0498f58-c99b-4fa8-a404-f7f43199c29f\") " Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.120953 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121017 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121004 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121046 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121050 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-node-log" (OuterVolumeSpecName: "node-log") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121083 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121099 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-slash" (OuterVolumeSpecName: "host-slash") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121170 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121229 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-cni-netd\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121293 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-cni-bin\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121152 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121366 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-node-log\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121395 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-slash\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121426 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml6f8\" (UniqueName: \"kubernetes.io/projected/1a796256-a1b4-41e5-b37a-b85c826684a1-kube-api-access-ml6f8\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121469 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-var-lib-openvswitch\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121498 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-log-socket\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121520 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1a796256-a1b4-41e5-b37a-b85c826684a1-ovnkube-config\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121549 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1a796256-a1b4-41e5-b37a-b85c826684a1-env-overrides\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121576 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-run-ovn\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121636 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-run-ovn-kubernetes\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121714 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121729 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-etc-openvswitch\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121774 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-systemd-units\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121807 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121839 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1a796256-a1b4-41e5-b37a-b85c826684a1-ovn-node-metrics-cert\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121874 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-run-openvswitch\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.121902 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-run-netns\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122023 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-kubelet\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122098 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1a796256-a1b4-41e5-b37a-b85c826684a1-ovnkube-script-lib\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122237 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-run-systemd\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122399 4943 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122427 4943 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122443 4943 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122487 4943 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122503 4943 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122518 4943 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122535 4943 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122553 4943 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122567 4943 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122580 4943 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122593 4943 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-slash\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122607 4943 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122621 4943 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-node-log\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122635 4943 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122648 4943 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-log-socket\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122661 4943 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.122678 4943 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.125979 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.128182 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0498f58-c99b-4fa8-a404-f7f43199c29f-kube-api-access-7wh9m" (OuterVolumeSpecName: "kube-api-access-7wh9m") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "kube-api-access-7wh9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.138296 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "d0498f58-c99b-4fa8-a404-f7f43199c29f" (UID: "d0498f58-c99b-4fa8-a404-f7f43199c29f"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223561 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-cni-bin\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223639 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-node-log\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223671 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-slash\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223690 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml6f8\" (UniqueName: \"kubernetes.io/projected/1a796256-a1b4-41e5-b37a-b85c826684a1-kube-api-access-ml6f8\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223708 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-var-lib-openvswitch\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223701 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-cni-bin\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223777 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-node-log\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223774 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-log-socket\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223727 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-log-socket\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223819 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-slash\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223834 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1a796256-a1b4-41e5-b37a-b85c826684a1-ovnkube-config\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223864 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1a796256-a1b4-41e5-b37a-b85c826684a1-env-overrides\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223887 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-run-ovn\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223919 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-run-ovn-kubernetes\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.223985 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-etc-openvswitch\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224009 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-systemd-units\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224042 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224080 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1a796256-a1b4-41e5-b37a-b85c826684a1-ovn-node-metrics-cert\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224118 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-run-openvswitch\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224141 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-run-netns\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224157 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-var-lib-openvswitch\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224174 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-kubelet\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224187 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-etc-openvswitch\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224234 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1a796256-a1b4-41e5-b37a-b85c826684a1-ovnkube-script-lib\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224257 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-run-systemd\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224304 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-cni-netd\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224358 4943 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d0498f58-c99b-4fa8-a404-f7f43199c29f-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224372 4943 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d0498f58-c99b-4fa8-a404-f7f43199c29f-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224384 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wh9m\" (UniqueName: \"kubernetes.io/projected/d0498f58-c99b-4fa8-a404-f7f43199c29f-kube-api-access-7wh9m\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224414 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-cni-netd\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224443 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-systemd-units\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224464 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.224852 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1a796256-a1b4-41e5-b37a-b85c826684a1-ovnkube-config\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.225174 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1a796256-a1b4-41e5-b37a-b85c826684a1-env-overrides\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.225235 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-run-ovn\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.225269 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-run-ovn-kubernetes\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.225300 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-kubelet\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.225328 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-run-openvswitch\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.225353 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-host-run-netns\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.225471 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1a796256-a1b4-41e5-b37a-b85c826684a1-run-systemd\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.226342 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1a796256-a1b4-41e5-b37a-b85c826684a1-ovnkube-script-lib\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.229458 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1a796256-a1b4-41e5-b37a-b85c826684a1-ovn-node-metrics-cert\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.242585 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml6f8\" (UniqueName: \"kubernetes.io/projected/1a796256-a1b4-41e5-b37a-b85c826684a1-kube-api-access-ml6f8\") pod \"ovnkube-node-lptpz\" (UID: \"1a796256-a1b4-41e5-b37a-b85c826684a1\") " pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.317689 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.734775 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-wkbf5_b9cb494f-7d4c-45f2-8b9b-e35c42b41c79/kube-multus/2.log" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.734891 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-wkbf5" event={"ID":"b9cb494f-7d4c-45f2-8b9b-e35c42b41c79","Type":"ContainerStarted","Data":"9a376837471528296694c7c39bdf00913e53e45f4d3aed359286f4f05623efd9"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.737052 4943 generic.go:334] "Generic (PLEG): container finished" podID="1a796256-a1b4-41e5-b37a-b85c826684a1" containerID="d066f4898bb87c2606de0e2e80779db19c8a9eb975d5941b64afb76ad2a0492a" exitCode=0 Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.737104 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" event={"ID":"1a796256-a1b4-41e5-b37a-b85c826684a1","Type":"ContainerDied","Data":"d066f4898bb87c2606de0e2e80779db19c8a9eb975d5941b64afb76ad2a0492a"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.737134 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" event={"ID":"1a796256-a1b4-41e5-b37a-b85c826684a1","Type":"ContainerStarted","Data":"0a06871ca22395edc45b4b10c104beeb31dbf3351d690ac3a52f94a626780861"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.740829 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovnkube-controller/3.log" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.744227 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovn-acl-logging/0.log" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.744940 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-7g4wf_d0498f58-c99b-4fa8-a404-f7f43199c29f/ovn-controller/0.log" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745430 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce" exitCode=0 Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745473 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284" exitCode=0 Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745490 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98" exitCode=0 Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745503 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744" exitCode=0 Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745516 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba" exitCode=0 Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745528 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761" exitCode=0 Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745540 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5" exitCode=143 Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745559 4943 generic.go:334] "Generic (PLEG): container finished" podID="d0498f58-c99b-4fa8-a404-f7f43199c29f" containerID="aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326" exitCode=143 Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745607 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745650 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745671 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745690 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745709 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745728 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745763 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745790 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745802 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745813 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745824 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745835 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745845 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745855 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745865 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745880 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745896 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745911 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745921 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745931 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745942 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745952 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745962 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745973 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745983 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.745993 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746006 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746022 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746034 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746045 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746056 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746066 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746076 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746086 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746096 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746108 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746118 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746132 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" event={"ID":"d0498f58-c99b-4fa8-a404-f7f43199c29f","Type":"ContainerDied","Data":"e9bbac1f390cd447ef4c261ccdc5cf69edafa1c4c5d23cae22c29d3289bf8d17"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746149 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746161 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746171 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746181 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746191 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746231 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746241 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746252 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746262 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746272 4943 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966"} Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746296 4943 scope.go:117] "RemoveContainer" containerID="68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.746521 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7g4wf" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.773463 4943 scope.go:117] "RemoveContainer" containerID="16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.799375 4943 scope.go:117] "RemoveContainer" containerID="1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.825286 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7g4wf"] Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.829759 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7g4wf"] Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.841592 4943 scope.go:117] "RemoveContainer" containerID="7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.857998 4943 scope.go:117] "RemoveContainer" containerID="86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.892347 4943 scope.go:117] "RemoveContainer" containerID="e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.909529 4943 scope.go:117] "RemoveContainer" containerID="ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.926991 4943 scope.go:117] "RemoveContainer" containerID="8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.945581 4943 scope.go:117] "RemoveContainer" containerID="aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326" Dec 04 10:29:28 crc kubenswrapper[4943]: I1204 10:29:28.969980 4943 scope.go:117] "RemoveContainer" containerID="2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.000045 4943 scope.go:117] "RemoveContainer" containerID="68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce" Dec 04 10:29:29 crc kubenswrapper[4943]: E1204 10:29:29.000579 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce\": container with ID starting with 68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce not found: ID does not exist" containerID="68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.000623 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce"} err="failed to get container status \"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce\": rpc error: code = NotFound desc = could not find container \"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce\": container with ID starting with 68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.000653 4943 scope.go:117] "RemoveContainer" containerID="16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a" Dec 04 10:29:29 crc kubenswrapper[4943]: E1204 10:29:29.001011 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a\": container with ID starting with 16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a not found: ID does not exist" containerID="16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.001034 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a"} err="failed to get container status \"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a\": rpc error: code = NotFound desc = could not find container \"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a\": container with ID starting with 16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.001062 4943 scope.go:117] "RemoveContainer" containerID="1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284" Dec 04 10:29:29 crc kubenswrapper[4943]: E1204 10:29:29.001440 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\": container with ID starting with 1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284 not found: ID does not exist" containerID="1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.001475 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284"} err="failed to get container status \"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\": rpc error: code = NotFound desc = could not find container \"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\": container with ID starting with 1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.001501 4943 scope.go:117] "RemoveContainer" containerID="7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98" Dec 04 10:29:29 crc kubenswrapper[4943]: E1204 10:29:29.001823 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\": container with ID starting with 7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98 not found: ID does not exist" containerID="7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.001848 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98"} err="failed to get container status \"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\": rpc error: code = NotFound desc = could not find container \"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\": container with ID starting with 7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.001866 4943 scope.go:117] "RemoveContainer" containerID="86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744" Dec 04 10:29:29 crc kubenswrapper[4943]: E1204 10:29:29.002415 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\": container with ID starting with 86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744 not found: ID does not exist" containerID="86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.002438 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744"} err="failed to get container status \"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\": rpc error: code = NotFound desc = could not find container \"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\": container with ID starting with 86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.002462 4943 scope.go:117] "RemoveContainer" containerID="e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba" Dec 04 10:29:29 crc kubenswrapper[4943]: E1204 10:29:29.002705 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\": container with ID starting with e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba not found: ID does not exist" containerID="e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.002728 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba"} err="failed to get container status \"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\": rpc error: code = NotFound desc = could not find container \"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\": container with ID starting with e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.002745 4943 scope.go:117] "RemoveContainer" containerID="ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761" Dec 04 10:29:29 crc kubenswrapper[4943]: E1204 10:29:29.003053 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\": container with ID starting with ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761 not found: ID does not exist" containerID="ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.003095 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761"} err="failed to get container status \"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\": rpc error: code = NotFound desc = could not find container \"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\": container with ID starting with ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.003119 4943 scope.go:117] "RemoveContainer" containerID="8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5" Dec 04 10:29:29 crc kubenswrapper[4943]: E1204 10:29:29.003417 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\": container with ID starting with 8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5 not found: ID does not exist" containerID="8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.003442 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5"} err="failed to get container status \"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\": rpc error: code = NotFound desc = could not find container \"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\": container with ID starting with 8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.003459 4943 scope.go:117] "RemoveContainer" containerID="aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326" Dec 04 10:29:29 crc kubenswrapper[4943]: E1204 10:29:29.003749 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\": container with ID starting with aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326 not found: ID does not exist" containerID="aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.003778 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326"} err="failed to get container status \"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\": rpc error: code = NotFound desc = could not find container \"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\": container with ID starting with aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.003795 4943 scope.go:117] "RemoveContainer" containerID="2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966" Dec 04 10:29:29 crc kubenswrapper[4943]: E1204 10:29:29.004024 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\": container with ID starting with 2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966 not found: ID does not exist" containerID="2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.004050 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966"} err="failed to get container status \"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\": rpc error: code = NotFound desc = could not find container \"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\": container with ID starting with 2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.004068 4943 scope.go:117] "RemoveContainer" containerID="68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.004339 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce"} err="failed to get container status \"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce\": rpc error: code = NotFound desc = could not find container \"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce\": container with ID starting with 68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.004409 4943 scope.go:117] "RemoveContainer" containerID="16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.004662 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a"} err="failed to get container status \"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a\": rpc error: code = NotFound desc = could not find container \"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a\": container with ID starting with 16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.004687 4943 scope.go:117] "RemoveContainer" containerID="1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.004938 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284"} err="failed to get container status \"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\": rpc error: code = NotFound desc = could not find container \"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\": container with ID starting with 1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.004965 4943 scope.go:117] "RemoveContainer" containerID="7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.005319 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98"} err="failed to get container status \"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\": rpc error: code = NotFound desc = could not find container \"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\": container with ID starting with 7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.005358 4943 scope.go:117] "RemoveContainer" containerID="86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.005995 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744"} err="failed to get container status \"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\": rpc error: code = NotFound desc = could not find container \"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\": container with ID starting with 86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.006017 4943 scope.go:117] "RemoveContainer" containerID="e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.006192 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba"} err="failed to get container status \"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\": rpc error: code = NotFound desc = could not find container \"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\": container with ID starting with e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.006289 4943 scope.go:117] "RemoveContainer" containerID="ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.006483 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761"} err="failed to get container status \"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\": rpc error: code = NotFound desc = could not find container \"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\": container with ID starting with ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.006500 4943 scope.go:117] "RemoveContainer" containerID="8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.006745 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5"} err="failed to get container status \"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\": rpc error: code = NotFound desc = could not find container \"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\": container with ID starting with 8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.006796 4943 scope.go:117] "RemoveContainer" containerID="aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.007223 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326"} err="failed to get container status \"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\": rpc error: code = NotFound desc = could not find container \"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\": container with ID starting with aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.007245 4943 scope.go:117] "RemoveContainer" containerID="2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.007519 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966"} err="failed to get container status \"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\": rpc error: code = NotFound desc = could not find container \"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\": container with ID starting with 2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.007568 4943 scope.go:117] "RemoveContainer" containerID="68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.008104 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce"} err="failed to get container status \"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce\": rpc error: code = NotFound desc = could not find container \"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce\": container with ID starting with 68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.008129 4943 scope.go:117] "RemoveContainer" containerID="16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.008620 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a"} err="failed to get container status \"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a\": rpc error: code = NotFound desc = could not find container \"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a\": container with ID starting with 16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.008647 4943 scope.go:117] "RemoveContainer" containerID="1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.009047 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284"} err="failed to get container status \"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\": rpc error: code = NotFound desc = could not find container \"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\": container with ID starting with 1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.009080 4943 scope.go:117] "RemoveContainer" containerID="7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.009357 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98"} err="failed to get container status \"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\": rpc error: code = NotFound desc = could not find container \"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\": container with ID starting with 7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.009379 4943 scope.go:117] "RemoveContainer" containerID="86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.009715 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744"} err="failed to get container status \"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\": rpc error: code = NotFound desc = could not find container \"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\": container with ID starting with 86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.009741 4943 scope.go:117] "RemoveContainer" containerID="e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.009971 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba"} err="failed to get container status \"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\": rpc error: code = NotFound desc = could not find container \"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\": container with ID starting with e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.009990 4943 scope.go:117] "RemoveContainer" containerID="ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.010317 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761"} err="failed to get container status \"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\": rpc error: code = NotFound desc = could not find container \"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\": container with ID starting with ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.010342 4943 scope.go:117] "RemoveContainer" containerID="8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.010766 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5"} err="failed to get container status \"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\": rpc error: code = NotFound desc = could not find container \"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\": container with ID starting with 8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.010792 4943 scope.go:117] "RemoveContainer" containerID="aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.011296 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326"} err="failed to get container status \"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\": rpc error: code = NotFound desc = could not find container \"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\": container with ID starting with aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.011331 4943 scope.go:117] "RemoveContainer" containerID="2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.011640 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966"} err="failed to get container status \"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\": rpc error: code = NotFound desc = could not find container \"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\": container with ID starting with 2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.011667 4943 scope.go:117] "RemoveContainer" containerID="68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.011974 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce"} err="failed to get container status \"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce\": rpc error: code = NotFound desc = could not find container \"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce\": container with ID starting with 68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.012015 4943 scope.go:117] "RemoveContainer" containerID="16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.012290 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a"} err="failed to get container status \"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a\": rpc error: code = NotFound desc = could not find container \"16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a\": container with ID starting with 16096b82ab66e3a98a02cfe532fbb5f92fcaf86a682c6f05428bdac9f091898a not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.012311 4943 scope.go:117] "RemoveContainer" containerID="1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.012617 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284"} err="failed to get container status \"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\": rpc error: code = NotFound desc = could not find container \"1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284\": container with ID starting with 1401e0bc5dfd6c25252a2a314be2b90d73b2df79e38e34e2d26ad15ea00af284 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.012637 4943 scope.go:117] "RemoveContainer" containerID="7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.012891 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98"} err="failed to get container status \"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\": rpc error: code = NotFound desc = could not find container \"7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98\": container with ID starting with 7eefee6e9b3c75e744042e2b9dc9f6c7a40fb0755de061b72f4732c31fffad98 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.012908 4943 scope.go:117] "RemoveContainer" containerID="86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.013181 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744"} err="failed to get container status \"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\": rpc error: code = NotFound desc = could not find container \"86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744\": container with ID starting with 86843ca70ab60bf221c463096cfa1aea68aa55b3076224a27457c3b4e4c72744 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.013241 4943 scope.go:117] "RemoveContainer" containerID="e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.013498 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba"} err="failed to get container status \"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\": rpc error: code = NotFound desc = could not find container \"e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba\": container with ID starting with e2daa56533e328eb5e4e5c21ade07ba19e1cae72b3b36f0ff2ccdb740de741ba not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.013526 4943 scope.go:117] "RemoveContainer" containerID="ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.013747 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761"} err="failed to get container status \"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\": rpc error: code = NotFound desc = could not find container \"ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761\": container with ID starting with ec02b76903e70fa54b988af40c31da706c667158b0ada92cc7357e6dd8658761 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.013768 4943 scope.go:117] "RemoveContainer" containerID="8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.013996 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5"} err="failed to get container status \"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\": rpc error: code = NotFound desc = could not find container \"8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5\": container with ID starting with 8607bbd7e108ce520fd9b15c3d6ac8b3fa762bdb74a9a8b670bb286c3a4565c5 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.014023 4943 scope.go:117] "RemoveContainer" containerID="aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.014254 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326"} err="failed to get container status \"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\": rpc error: code = NotFound desc = could not find container \"aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326\": container with ID starting with aba71b977b05350241c56d0e69ead66054a252978706e7320a8181424f186326 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.014279 4943 scope.go:117] "RemoveContainer" containerID="2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.014497 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966"} err="failed to get container status \"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\": rpc error: code = NotFound desc = could not find container \"2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966\": container with ID starting with 2f0349df1f761c7c3aadf54b5026dd31cf46cd35ac198f0f150aac53c62d9966 not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.014518 4943 scope.go:117] "RemoveContainer" containerID="68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.014747 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce"} err="failed to get container status \"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce\": rpc error: code = NotFound desc = could not find container \"68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce\": container with ID starting with 68e028f5a42e4bcfe1f02fc61856900967d6a3233033308ff3ca7367b5d2abce not found: ID does not exist" Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.765281 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" event={"ID":"1a796256-a1b4-41e5-b37a-b85c826684a1","Type":"ContainerStarted","Data":"39d77bd34b3122aa5b0ae418e8130b097d5ab248b2fd8e68fb02cf50bd7d8ca2"} Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.765593 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" event={"ID":"1a796256-a1b4-41e5-b37a-b85c826684a1","Type":"ContainerStarted","Data":"df906bb6842becb6433078e6fe124a557f1236325e2ae2932416f673e1f0ead6"} Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.765610 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" event={"ID":"1a796256-a1b4-41e5-b37a-b85c826684a1","Type":"ContainerStarted","Data":"1efc2269428160d462058730c251102346cee66ef131c44217c91b79210e8769"} Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.765619 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" event={"ID":"1a796256-a1b4-41e5-b37a-b85c826684a1","Type":"ContainerStarted","Data":"36fe3ecca195063d2b8fa927950ac25b1428434d0a99b8c1cba3ae34b17eb41c"} Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.765627 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" event={"ID":"1a796256-a1b4-41e5-b37a-b85c826684a1","Type":"ContainerStarted","Data":"cf837fb0cd9fe3e2b7e5c61bd4a2ec13fd700e191929e3e479d76478a2ae2eaa"} Dec 04 10:29:29 crc kubenswrapper[4943]: I1204 10:29:29.765635 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" event={"ID":"1a796256-a1b4-41e5-b37a-b85c826684a1","Type":"ContainerStarted","Data":"5d021deb220d02476b392f640706e41d336a058cc2034d8838a9fb33bb30e540"} Dec 04 10:29:30 crc kubenswrapper[4943]: I1204 10:29:30.576441 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0498f58-c99b-4fa8-a404-f7f43199c29f" path="/var/lib/kubelet/pods/d0498f58-c99b-4fa8-a404-f7f43199c29f/volumes" Dec 04 10:29:31 crc kubenswrapper[4943]: I1204 10:29:31.786961 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" event={"ID":"1a796256-a1b4-41e5-b37a-b85c826684a1","Type":"ContainerStarted","Data":"4c59d98f175e5dcb96298d7d544a5959c52153eb1ecd0a005021d535905448df"} Dec 04 10:29:34 crc kubenswrapper[4943]: I1204 10:29:34.808448 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" event={"ID":"1a796256-a1b4-41e5-b37a-b85c826684a1","Type":"ContainerStarted","Data":"fd40e2040e9d2b49b45986f9bea04d3b773511509159ec38bf6a0826d5cdaf0e"} Dec 04 10:29:34 crc kubenswrapper[4943]: I1204 10:29:34.809120 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:34 crc kubenswrapper[4943]: I1204 10:29:34.809142 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:34 crc kubenswrapper[4943]: I1204 10:29:34.809156 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:34 crc kubenswrapper[4943]: I1204 10:29:34.835943 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:34 crc kubenswrapper[4943]: I1204 10:29:34.843577 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" podStartSLOduration=7.843532619 podStartE2EDuration="7.843532619s" podCreationTimestamp="2025-12-04 10:29:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:29:34.840871487 +0000 UTC m=+823.429847365" watchObservedRunningTime="2025-12-04 10:29:34.843532619 +0000 UTC m=+823.432508527" Dec 04 10:29:34 crc kubenswrapper[4943]: I1204 10:29:34.844560 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:52 crc kubenswrapper[4943]: I1204 10:29:52.973715 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw"] Dec 04 10:29:52 crc kubenswrapper[4943]: I1204 10:29:52.975578 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:29:52 crc kubenswrapper[4943]: I1204 10:29:52.979215 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 10:29:52 crc kubenswrapper[4943]: I1204 10:29:52.979867 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw"] Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.060250 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw\" (UID: \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.060327 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wdf4\" (UniqueName: \"kubernetes.io/projected/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-kube-api-access-7wdf4\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw\" (UID: \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.060484 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw\" (UID: \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.162574 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wdf4\" (UniqueName: \"kubernetes.io/projected/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-kube-api-access-7wdf4\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw\" (UID: \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.162669 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw\" (UID: \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.162733 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw\" (UID: \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.163233 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw\" (UID: \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.163282 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw\" (UID: \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.193436 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wdf4\" (UniqueName: \"kubernetes.io/projected/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-kube-api-access-7wdf4\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw\" (UID: \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.295547 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.534489 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw"] Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.928893 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" event={"ID":"7a3629de-83c4-4c6a-96e1-58d9bf06b57f","Type":"ContainerStarted","Data":"98a5f9f9773d2e05c62ab7ddef3054a3c74d1a180c029e12f4deaa25eb8d4b1c"} Dec 04 10:29:53 crc kubenswrapper[4943]: I1204 10:29:53.929379 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" event={"ID":"7a3629de-83c4-4c6a-96e1-58d9bf06b57f","Type":"ContainerStarted","Data":"df3235a715ff733dd4a63a52e6c3c6db8a761600e6628db0690e2367581b9525"} Dec 04 10:29:54 crc kubenswrapper[4943]: I1204 10:29:54.936730 4943 generic.go:334] "Generic (PLEG): container finished" podID="7a3629de-83c4-4c6a-96e1-58d9bf06b57f" containerID="98a5f9f9773d2e05c62ab7ddef3054a3c74d1a180c029e12f4deaa25eb8d4b1c" exitCode=0 Dec 04 10:29:54 crc kubenswrapper[4943]: I1204 10:29:54.936784 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" event={"ID":"7a3629de-83c4-4c6a-96e1-58d9bf06b57f","Type":"ContainerDied","Data":"98a5f9f9773d2e05c62ab7ddef3054a3c74d1a180c029e12f4deaa25eb8d4b1c"} Dec 04 10:29:54 crc kubenswrapper[4943]: I1204 10:29:54.939588 4943 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.309144 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-brgqp"] Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.311352 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.318021 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-brgqp"] Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.392771 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50295603-ccc5-48e9-8083-9094c05208d4-catalog-content\") pod \"redhat-operators-brgqp\" (UID: \"50295603-ccc5-48e9-8083-9094c05208d4\") " pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.392823 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50295603-ccc5-48e9-8083-9094c05208d4-utilities\") pod \"redhat-operators-brgqp\" (UID: \"50295603-ccc5-48e9-8083-9094c05208d4\") " pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.392889 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2q9s\" (UniqueName: \"kubernetes.io/projected/50295603-ccc5-48e9-8083-9094c05208d4-kube-api-access-q2q9s\") pod \"redhat-operators-brgqp\" (UID: \"50295603-ccc5-48e9-8083-9094c05208d4\") " pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.494381 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50295603-ccc5-48e9-8083-9094c05208d4-catalog-content\") pod \"redhat-operators-brgqp\" (UID: \"50295603-ccc5-48e9-8083-9094c05208d4\") " pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.494446 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50295603-ccc5-48e9-8083-9094c05208d4-utilities\") pod \"redhat-operators-brgqp\" (UID: \"50295603-ccc5-48e9-8083-9094c05208d4\") " pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.494472 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2q9s\" (UniqueName: \"kubernetes.io/projected/50295603-ccc5-48e9-8083-9094c05208d4-kube-api-access-q2q9s\") pod \"redhat-operators-brgqp\" (UID: \"50295603-ccc5-48e9-8083-9094c05208d4\") " pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.494944 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50295603-ccc5-48e9-8083-9094c05208d4-utilities\") pod \"redhat-operators-brgqp\" (UID: \"50295603-ccc5-48e9-8083-9094c05208d4\") " pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.494984 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50295603-ccc5-48e9-8083-9094c05208d4-catalog-content\") pod \"redhat-operators-brgqp\" (UID: \"50295603-ccc5-48e9-8083-9094c05208d4\") " pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.513368 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2q9s\" (UniqueName: \"kubernetes.io/projected/50295603-ccc5-48e9-8083-9094c05208d4-kube-api-access-q2q9s\") pod \"redhat-operators-brgqp\" (UID: \"50295603-ccc5-48e9-8083-9094c05208d4\") " pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.625797 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.884238 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-brgqp"] Dec 04 10:29:55 crc kubenswrapper[4943]: I1204 10:29:55.942787 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brgqp" event={"ID":"50295603-ccc5-48e9-8083-9094c05208d4","Type":"ContainerStarted","Data":"c294a264945e4387f8488a2820a39f227b57deb25f0f31b7a778edf631854dd5"} Dec 04 10:29:56 crc kubenswrapper[4943]: I1204 10:29:56.950859 4943 generic.go:334] "Generic (PLEG): container finished" podID="50295603-ccc5-48e9-8083-9094c05208d4" containerID="5ec91c4f3181aca0743ae96823e2140c8c68aa14e34bcab47228fbaa85b13990" exitCode=0 Dec 04 10:29:56 crc kubenswrapper[4943]: I1204 10:29:56.950951 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brgqp" event={"ID":"50295603-ccc5-48e9-8083-9094c05208d4","Type":"ContainerDied","Data":"5ec91c4f3181aca0743ae96823e2140c8c68aa14e34bcab47228fbaa85b13990"} Dec 04 10:29:56 crc kubenswrapper[4943]: I1204 10:29:56.953658 4943 generic.go:334] "Generic (PLEG): container finished" podID="7a3629de-83c4-4c6a-96e1-58d9bf06b57f" containerID="5b0953dac9cd4e12f61ee43d89ebc1b5f8a9022f9e2ca8c6ab71ebf75370b833" exitCode=0 Dec 04 10:29:56 crc kubenswrapper[4943]: I1204 10:29:56.953709 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" event={"ID":"7a3629de-83c4-4c6a-96e1-58d9bf06b57f","Type":"ContainerDied","Data":"5b0953dac9cd4e12f61ee43d89ebc1b5f8a9022f9e2ca8c6ab71ebf75370b833"} Dec 04 10:29:57 crc kubenswrapper[4943]: I1204 10:29:57.962079 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brgqp" event={"ID":"50295603-ccc5-48e9-8083-9094c05208d4","Type":"ContainerStarted","Data":"1b314edf98eaeeb8d228ef5760e610d994c75c1af677bb592ac13aa9a1a542ed"} Dec 04 10:29:57 crc kubenswrapper[4943]: I1204 10:29:57.966821 4943 generic.go:334] "Generic (PLEG): container finished" podID="7a3629de-83c4-4c6a-96e1-58d9bf06b57f" containerID="c3fd40daaa3a2df5f64aa4b97e270637ee3a85475ba08aa38e0e42839962fdcb" exitCode=0 Dec 04 10:29:57 crc kubenswrapper[4943]: I1204 10:29:57.966854 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" event={"ID":"7a3629de-83c4-4c6a-96e1-58d9bf06b57f","Type":"ContainerDied","Data":"c3fd40daaa3a2df5f64aa4b97e270637ee3a85475ba08aa38e0e42839962fdcb"} Dec 04 10:29:58 crc kubenswrapper[4943]: I1204 10:29:58.355925 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lptpz" Dec 04 10:29:58 crc kubenswrapper[4943]: I1204 10:29:58.976101 4943 generic.go:334] "Generic (PLEG): container finished" podID="50295603-ccc5-48e9-8083-9094c05208d4" containerID="1b314edf98eaeeb8d228ef5760e610d994c75c1af677bb592ac13aa9a1a542ed" exitCode=0 Dec 04 10:29:58 crc kubenswrapper[4943]: I1204 10:29:58.977178 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brgqp" event={"ID":"50295603-ccc5-48e9-8083-9094c05208d4","Type":"ContainerDied","Data":"1b314edf98eaeeb8d228ef5760e610d994c75c1af677bb592ac13aa9a1a542ed"} Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.262549 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.359516 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wdf4\" (UniqueName: \"kubernetes.io/projected/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-kube-api-access-7wdf4\") pod \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\" (UID: \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\") " Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.359616 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-bundle\") pod \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\" (UID: \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\") " Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.359703 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-util\") pod \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\" (UID: \"7a3629de-83c4-4c6a-96e1-58d9bf06b57f\") " Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.361424 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-bundle" (OuterVolumeSpecName: "bundle") pod "7a3629de-83c4-4c6a-96e1-58d9bf06b57f" (UID: "7a3629de-83c4-4c6a-96e1-58d9bf06b57f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.366679 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-kube-api-access-7wdf4" (OuterVolumeSpecName: "kube-api-access-7wdf4") pod "7a3629de-83c4-4c6a-96e1-58d9bf06b57f" (UID: "7a3629de-83c4-4c6a-96e1-58d9bf06b57f"). InnerVolumeSpecName "kube-api-access-7wdf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.373465 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-util" (OuterVolumeSpecName: "util") pod "7a3629de-83c4-4c6a-96e1-58d9bf06b57f" (UID: "7a3629de-83c4-4c6a-96e1-58d9bf06b57f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.462558 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.462675 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-util\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.462754 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wdf4\" (UniqueName: \"kubernetes.io/projected/7a3629de-83c4-4c6a-96e1-58d9bf06b57f-kube-api-access-7wdf4\") on node \"crc\" DevicePath \"\"" Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.987444 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" event={"ID":"7a3629de-83c4-4c6a-96e1-58d9bf06b57f","Type":"ContainerDied","Data":"df3235a715ff733dd4a63a52e6c3c6db8a761600e6628db0690e2367581b9525"} Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.987491 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df3235a715ff733dd4a63a52e6c3c6db8a761600e6628db0690e2367581b9525" Dec 04 10:29:59 crc kubenswrapper[4943]: I1204 10:29:59.987543 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.159475 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc"] Dec 04 10:30:00 crc kubenswrapper[4943]: E1204 10:30:00.160045 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a3629de-83c4-4c6a-96e1-58d9bf06b57f" containerName="util" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.160062 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a3629de-83c4-4c6a-96e1-58d9bf06b57f" containerName="util" Dec 04 10:30:00 crc kubenswrapper[4943]: E1204 10:30:00.160083 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a3629de-83c4-4c6a-96e1-58d9bf06b57f" containerName="pull" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.160090 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a3629de-83c4-4c6a-96e1-58d9bf06b57f" containerName="pull" Dec 04 10:30:00 crc kubenswrapper[4943]: E1204 10:30:00.160106 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a3629de-83c4-4c6a-96e1-58d9bf06b57f" containerName="extract" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.160114 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a3629de-83c4-4c6a-96e1-58d9bf06b57f" containerName="extract" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.160249 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a3629de-83c4-4c6a-96e1-58d9bf06b57f" containerName="extract" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.160729 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.163057 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.163553 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.171466 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc"] Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.281230 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktx2r\" (UniqueName: \"kubernetes.io/projected/de2dd00b-d575-473b-9d81-1d7242955ee9-kube-api-access-ktx2r\") pod \"collect-profiles-29414070-ww8cc\" (UID: \"de2dd00b-d575-473b-9d81-1d7242955ee9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.281318 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de2dd00b-d575-473b-9d81-1d7242955ee9-config-volume\") pod \"collect-profiles-29414070-ww8cc\" (UID: \"de2dd00b-d575-473b-9d81-1d7242955ee9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.281409 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de2dd00b-d575-473b-9d81-1d7242955ee9-secret-volume\") pod \"collect-profiles-29414070-ww8cc\" (UID: \"de2dd00b-d575-473b-9d81-1d7242955ee9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.382429 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de2dd00b-d575-473b-9d81-1d7242955ee9-secret-volume\") pod \"collect-profiles-29414070-ww8cc\" (UID: \"de2dd00b-d575-473b-9d81-1d7242955ee9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.382564 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktx2r\" (UniqueName: \"kubernetes.io/projected/de2dd00b-d575-473b-9d81-1d7242955ee9-kube-api-access-ktx2r\") pod \"collect-profiles-29414070-ww8cc\" (UID: \"de2dd00b-d575-473b-9d81-1d7242955ee9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.382589 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de2dd00b-d575-473b-9d81-1d7242955ee9-config-volume\") pod \"collect-profiles-29414070-ww8cc\" (UID: \"de2dd00b-d575-473b-9d81-1d7242955ee9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.384428 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de2dd00b-d575-473b-9d81-1d7242955ee9-config-volume\") pod \"collect-profiles-29414070-ww8cc\" (UID: \"de2dd00b-d575-473b-9d81-1d7242955ee9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.387449 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de2dd00b-d575-473b-9d81-1d7242955ee9-secret-volume\") pod \"collect-profiles-29414070-ww8cc\" (UID: \"de2dd00b-d575-473b-9d81-1d7242955ee9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.401888 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktx2r\" (UniqueName: \"kubernetes.io/projected/de2dd00b-d575-473b-9d81-1d7242955ee9-kube-api-access-ktx2r\") pod \"collect-profiles-29414070-ww8cc\" (UID: \"de2dd00b-d575-473b-9d81-1d7242955ee9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.489955 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:00 crc kubenswrapper[4943]: W1204 10:30:00.889722 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde2dd00b_d575_473b_9d81_1d7242955ee9.slice/crio-1b7ca87fb33e1ded9f587a7b24b0d763b54eb138027bb9ee2ca7596bc0ee4a23 WatchSource:0}: Error finding container 1b7ca87fb33e1ded9f587a7b24b0d763b54eb138027bb9ee2ca7596bc0ee4a23: Status 404 returned error can't find the container with id 1b7ca87fb33e1ded9f587a7b24b0d763b54eb138027bb9ee2ca7596bc0ee4a23 Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.897223 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc"] Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.996709 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brgqp" event={"ID":"50295603-ccc5-48e9-8083-9094c05208d4","Type":"ContainerStarted","Data":"4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e"} Dec 04 10:30:00 crc kubenswrapper[4943]: I1204 10:30:00.997832 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" event={"ID":"de2dd00b-d575-473b-9d81-1d7242955ee9","Type":"ContainerStarted","Data":"1b7ca87fb33e1ded9f587a7b24b0d763b54eb138027bb9ee2ca7596bc0ee4a23"} Dec 04 10:30:02 crc kubenswrapper[4943]: I1204 10:30:02.004508 4943 generic.go:334] "Generic (PLEG): container finished" podID="de2dd00b-d575-473b-9d81-1d7242955ee9" containerID="45f61b632f8500fdc15f35cd0ab258c190d646810bda35c5c2400e359bd0f686" exitCode=0 Dec 04 10:30:02 crc kubenswrapper[4943]: I1204 10:30:02.004588 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" event={"ID":"de2dd00b-d575-473b-9d81-1d7242955ee9","Type":"ContainerDied","Data":"45f61b632f8500fdc15f35cd0ab258c190d646810bda35c5c2400e359bd0f686"} Dec 04 10:30:02 crc kubenswrapper[4943]: I1204 10:30:02.023998 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-brgqp" podStartSLOduration=3.996073867 podStartE2EDuration="7.023975938s" podCreationTimestamp="2025-12-04 10:29:55 +0000 UTC" firstStartedPulling="2025-12-04 10:29:56.953381539 +0000 UTC m=+845.542357447" lastFinishedPulling="2025-12-04 10:29:59.98128364 +0000 UTC m=+848.570259518" observedRunningTime="2025-12-04 10:30:01.015125034 +0000 UTC m=+849.604100912" watchObservedRunningTime="2025-12-04 10:30:02.023975938 +0000 UTC m=+850.612951806" Dec 04 10:30:03 crc kubenswrapper[4943]: I1204 10:30:03.280069 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:03 crc kubenswrapper[4943]: I1204 10:30:03.429141 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktx2r\" (UniqueName: \"kubernetes.io/projected/de2dd00b-d575-473b-9d81-1d7242955ee9-kube-api-access-ktx2r\") pod \"de2dd00b-d575-473b-9d81-1d7242955ee9\" (UID: \"de2dd00b-d575-473b-9d81-1d7242955ee9\") " Dec 04 10:30:03 crc kubenswrapper[4943]: I1204 10:30:03.429226 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de2dd00b-d575-473b-9d81-1d7242955ee9-secret-volume\") pod \"de2dd00b-d575-473b-9d81-1d7242955ee9\" (UID: \"de2dd00b-d575-473b-9d81-1d7242955ee9\") " Dec 04 10:30:03 crc kubenswrapper[4943]: I1204 10:30:03.429342 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de2dd00b-d575-473b-9d81-1d7242955ee9-config-volume\") pod \"de2dd00b-d575-473b-9d81-1d7242955ee9\" (UID: \"de2dd00b-d575-473b-9d81-1d7242955ee9\") " Dec 04 10:30:03 crc kubenswrapper[4943]: I1204 10:30:03.430168 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de2dd00b-d575-473b-9d81-1d7242955ee9-config-volume" (OuterVolumeSpecName: "config-volume") pod "de2dd00b-d575-473b-9d81-1d7242955ee9" (UID: "de2dd00b-d575-473b-9d81-1d7242955ee9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:30:03 crc kubenswrapper[4943]: I1204 10:30:03.436518 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de2dd00b-d575-473b-9d81-1d7242955ee9-kube-api-access-ktx2r" (OuterVolumeSpecName: "kube-api-access-ktx2r") pod "de2dd00b-d575-473b-9d81-1d7242955ee9" (UID: "de2dd00b-d575-473b-9d81-1d7242955ee9"). InnerVolumeSpecName "kube-api-access-ktx2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:30:03 crc kubenswrapper[4943]: I1204 10:30:03.437426 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2dd00b-d575-473b-9d81-1d7242955ee9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "de2dd00b-d575-473b-9d81-1d7242955ee9" (UID: "de2dd00b-d575-473b-9d81-1d7242955ee9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:30:03 crc kubenswrapper[4943]: I1204 10:30:03.531101 4943 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de2dd00b-d575-473b-9d81-1d7242955ee9-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 10:30:03 crc kubenswrapper[4943]: I1204 10:30:03.531136 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktx2r\" (UniqueName: \"kubernetes.io/projected/de2dd00b-d575-473b-9d81-1d7242955ee9-kube-api-access-ktx2r\") on node \"crc\" DevicePath \"\"" Dec 04 10:30:03 crc kubenswrapper[4943]: I1204 10:30:03.531152 4943 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de2dd00b-d575-473b-9d81-1d7242955ee9-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 10:30:04 crc kubenswrapper[4943]: I1204 10:30:04.016930 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" event={"ID":"de2dd00b-d575-473b-9d81-1d7242955ee9","Type":"ContainerDied","Data":"1b7ca87fb33e1ded9f587a7b24b0d763b54eb138027bb9ee2ca7596bc0ee4a23"} Dec 04 10:30:04 crc kubenswrapper[4943]: I1204 10:30:04.016969 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414070-ww8cc" Dec 04 10:30:04 crc kubenswrapper[4943]: I1204 10:30:04.016989 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b7ca87fb33e1ded9f587a7b24b0d763b54eb138027bb9ee2ca7596bc0ee4a23" Dec 04 10:30:05 crc kubenswrapper[4943]: I1204 10:30:05.626969 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:30:05 crc kubenswrapper[4943]: I1204 10:30:05.627458 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:30:05 crc kubenswrapper[4943]: I1204 10:30:05.663927 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:30:06 crc kubenswrapper[4943]: I1204 10:30:06.063022 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:30:07 crc kubenswrapper[4943]: I1204 10:30:07.297043 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-brgqp"] Dec 04 10:30:08 crc kubenswrapper[4943]: I1204 10:30:08.037340 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-brgqp" podUID="50295603-ccc5-48e9-8083-9094c05208d4" containerName="registry-server" containerID="cri-o://4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e" gracePeriod=2 Dec 04 10:30:08 crc kubenswrapper[4943]: I1204 10:30:08.359750 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:30:08 crc kubenswrapper[4943]: I1204 10:30:08.492192 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2q9s\" (UniqueName: \"kubernetes.io/projected/50295603-ccc5-48e9-8083-9094c05208d4-kube-api-access-q2q9s\") pod \"50295603-ccc5-48e9-8083-9094c05208d4\" (UID: \"50295603-ccc5-48e9-8083-9094c05208d4\") " Dec 04 10:30:08 crc kubenswrapper[4943]: I1204 10:30:08.492328 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50295603-ccc5-48e9-8083-9094c05208d4-catalog-content\") pod \"50295603-ccc5-48e9-8083-9094c05208d4\" (UID: \"50295603-ccc5-48e9-8083-9094c05208d4\") " Dec 04 10:30:08 crc kubenswrapper[4943]: I1204 10:30:08.492432 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50295603-ccc5-48e9-8083-9094c05208d4-utilities\") pod \"50295603-ccc5-48e9-8083-9094c05208d4\" (UID: \"50295603-ccc5-48e9-8083-9094c05208d4\") " Dec 04 10:30:08 crc kubenswrapper[4943]: I1204 10:30:08.493303 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50295603-ccc5-48e9-8083-9094c05208d4-utilities" (OuterVolumeSpecName: "utilities") pod "50295603-ccc5-48e9-8083-9094c05208d4" (UID: "50295603-ccc5-48e9-8083-9094c05208d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:30:08 crc kubenswrapper[4943]: I1204 10:30:08.498033 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50295603-ccc5-48e9-8083-9094c05208d4-kube-api-access-q2q9s" (OuterVolumeSpecName: "kube-api-access-q2q9s") pod "50295603-ccc5-48e9-8083-9094c05208d4" (UID: "50295603-ccc5-48e9-8083-9094c05208d4"). InnerVolumeSpecName "kube-api-access-q2q9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:30:08 crc kubenswrapper[4943]: I1204 10:30:08.593565 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50295603-ccc5-48e9-8083-9094c05208d4-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:30:08 crc kubenswrapper[4943]: I1204 10:30:08.593621 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2q9s\" (UniqueName: \"kubernetes.io/projected/50295603-ccc5-48e9-8083-9094c05208d4-kube-api-access-q2q9s\") on node \"crc\" DevicePath \"\"" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.043162 4943 generic.go:334] "Generic (PLEG): container finished" podID="50295603-ccc5-48e9-8083-9094c05208d4" containerID="4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e" exitCode=0 Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.043212 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-brgqp" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.043232 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brgqp" event={"ID":"50295603-ccc5-48e9-8083-9094c05208d4","Type":"ContainerDied","Data":"4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e"} Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.043278 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-brgqp" event={"ID":"50295603-ccc5-48e9-8083-9094c05208d4","Type":"ContainerDied","Data":"c294a264945e4387f8488a2820a39f227b57deb25f0f31b7a778edf631854dd5"} Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.043367 4943 scope.go:117] "RemoveContainer" containerID="4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.059391 4943 scope.go:117] "RemoveContainer" containerID="1b314edf98eaeeb8d228ef5760e610d994c75c1af677bb592ac13aa9a1a542ed" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.076713 4943 scope.go:117] "RemoveContainer" containerID="5ec91c4f3181aca0743ae96823e2140c8c68aa14e34bcab47228fbaa85b13990" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.094057 4943 scope.go:117] "RemoveContainer" containerID="4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e" Dec 04 10:30:09 crc kubenswrapper[4943]: E1204 10:30:09.094544 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e\": container with ID starting with 4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e not found: ID does not exist" containerID="4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.094588 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e"} err="failed to get container status \"4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e\": rpc error: code = NotFound desc = could not find container \"4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e\": container with ID starting with 4cda10c0c04bed8408c1f5597ef8d16e2017bb3ad700b7a3c59e8a51eec3003e not found: ID does not exist" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.094617 4943 scope.go:117] "RemoveContainer" containerID="1b314edf98eaeeb8d228ef5760e610d994c75c1af677bb592ac13aa9a1a542ed" Dec 04 10:30:09 crc kubenswrapper[4943]: E1204 10:30:09.094886 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b314edf98eaeeb8d228ef5760e610d994c75c1af677bb592ac13aa9a1a542ed\": container with ID starting with 1b314edf98eaeeb8d228ef5760e610d994c75c1af677bb592ac13aa9a1a542ed not found: ID does not exist" containerID="1b314edf98eaeeb8d228ef5760e610d994c75c1af677bb592ac13aa9a1a542ed" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.094908 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b314edf98eaeeb8d228ef5760e610d994c75c1af677bb592ac13aa9a1a542ed"} err="failed to get container status \"1b314edf98eaeeb8d228ef5760e610d994c75c1af677bb592ac13aa9a1a542ed\": rpc error: code = NotFound desc = could not find container \"1b314edf98eaeeb8d228ef5760e610d994c75c1af677bb592ac13aa9a1a542ed\": container with ID starting with 1b314edf98eaeeb8d228ef5760e610d994c75c1af677bb592ac13aa9a1a542ed not found: ID does not exist" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.094931 4943 scope.go:117] "RemoveContainer" containerID="5ec91c4f3181aca0743ae96823e2140c8c68aa14e34bcab47228fbaa85b13990" Dec 04 10:30:09 crc kubenswrapper[4943]: E1204 10:30:09.095475 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ec91c4f3181aca0743ae96823e2140c8c68aa14e34bcab47228fbaa85b13990\": container with ID starting with 5ec91c4f3181aca0743ae96823e2140c8c68aa14e34bcab47228fbaa85b13990 not found: ID does not exist" containerID="5ec91c4f3181aca0743ae96823e2140c8c68aa14e34bcab47228fbaa85b13990" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.095497 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ec91c4f3181aca0743ae96823e2140c8c68aa14e34bcab47228fbaa85b13990"} err="failed to get container status \"5ec91c4f3181aca0743ae96823e2140c8c68aa14e34bcab47228fbaa85b13990\": rpc error: code = NotFound desc = could not find container \"5ec91c4f3181aca0743ae96823e2140c8c68aa14e34bcab47228fbaa85b13990\": container with ID starting with 5ec91c4f3181aca0743ae96823e2140c8c68aa14e34bcab47228fbaa85b13990 not found: ID does not exist" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.590154 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50295603-ccc5-48e9-8083-9094c05208d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50295603-ccc5-48e9-8083-9094c05208d4" (UID: "50295603-ccc5-48e9-8083-9094c05208d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.607299 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50295603-ccc5-48e9-8083-9094c05208d4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.670928 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-brgqp"] Dec 04 10:30:09 crc kubenswrapper[4943]: I1204 10:30:09.681975 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-brgqp"] Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.395296 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls"] Dec 04 10:30:10 crc kubenswrapper[4943]: E1204 10:30:10.395908 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2dd00b-d575-473b-9d81-1d7242955ee9" containerName="collect-profiles" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.395941 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2dd00b-d575-473b-9d81-1d7242955ee9" containerName="collect-profiles" Dec 04 10:30:10 crc kubenswrapper[4943]: E1204 10:30:10.395954 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50295603-ccc5-48e9-8083-9094c05208d4" containerName="extract-content" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.395962 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="50295603-ccc5-48e9-8083-9094c05208d4" containerName="extract-content" Dec 04 10:30:10 crc kubenswrapper[4943]: E1204 10:30:10.395980 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50295603-ccc5-48e9-8083-9094c05208d4" containerName="registry-server" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.395989 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="50295603-ccc5-48e9-8083-9094c05208d4" containerName="registry-server" Dec 04 10:30:10 crc kubenswrapper[4943]: E1204 10:30:10.396004 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50295603-ccc5-48e9-8083-9094c05208d4" containerName="extract-utilities" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.396012 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="50295603-ccc5-48e9-8083-9094c05208d4" containerName="extract-utilities" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.396167 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="de2dd00b-d575-473b-9d81-1d7242955ee9" containerName="collect-profiles" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.396187 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="50295603-ccc5-48e9-8083-9094c05208d4" containerName="registry-server" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.396819 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.399119 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-w8q2w" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.399577 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.411132 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.411138 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.411316 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.416792 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls"] Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.519388 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0ca89ccd-b04f-480d-9a4e-a96b534b3ca6-apiservice-cert\") pod \"metallb-operator-controller-manager-5697dbbbd-r8zls\" (UID: \"0ca89ccd-b04f-480d-9a4e-a96b534b3ca6\") " pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.519488 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0ca89ccd-b04f-480d-9a4e-a96b534b3ca6-webhook-cert\") pod \"metallb-operator-controller-manager-5697dbbbd-r8zls\" (UID: \"0ca89ccd-b04f-480d-9a4e-a96b534b3ca6\") " pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.519521 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6rbj\" (UniqueName: \"kubernetes.io/projected/0ca89ccd-b04f-480d-9a4e-a96b534b3ca6-kube-api-access-d6rbj\") pod \"metallb-operator-controller-manager-5697dbbbd-r8zls\" (UID: \"0ca89ccd-b04f-480d-9a4e-a96b534b3ca6\") " pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.571720 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50295603-ccc5-48e9-8083-9094c05208d4" path="/var/lib/kubelet/pods/50295603-ccc5-48e9-8083-9094c05208d4/volumes" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.620832 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0ca89ccd-b04f-480d-9a4e-a96b534b3ca6-apiservice-cert\") pod \"metallb-operator-controller-manager-5697dbbbd-r8zls\" (UID: \"0ca89ccd-b04f-480d-9a4e-a96b534b3ca6\") " pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.620918 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0ca89ccd-b04f-480d-9a4e-a96b534b3ca6-webhook-cert\") pod \"metallb-operator-controller-manager-5697dbbbd-r8zls\" (UID: \"0ca89ccd-b04f-480d-9a4e-a96b534b3ca6\") " pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.620939 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6rbj\" (UniqueName: \"kubernetes.io/projected/0ca89ccd-b04f-480d-9a4e-a96b534b3ca6-kube-api-access-d6rbj\") pod \"metallb-operator-controller-manager-5697dbbbd-r8zls\" (UID: \"0ca89ccd-b04f-480d-9a4e-a96b534b3ca6\") " pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.626222 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0ca89ccd-b04f-480d-9a4e-a96b534b3ca6-apiservice-cert\") pod \"metallb-operator-controller-manager-5697dbbbd-r8zls\" (UID: \"0ca89ccd-b04f-480d-9a4e-a96b534b3ca6\") " pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.626686 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0ca89ccd-b04f-480d-9a4e-a96b534b3ca6-webhook-cert\") pod \"metallb-operator-controller-manager-5697dbbbd-r8zls\" (UID: \"0ca89ccd-b04f-480d-9a4e-a96b534b3ca6\") " pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.639308 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6rbj\" (UniqueName: \"kubernetes.io/projected/0ca89ccd-b04f-480d-9a4e-a96b534b3ca6-kube-api-access-d6rbj\") pod \"metallb-operator-controller-manager-5697dbbbd-r8zls\" (UID: \"0ca89ccd-b04f-480d-9a4e-a96b534b3ca6\") " pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.655018 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt"] Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.655735 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.657718 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.658602 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.661077 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-s7n8n" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.711856 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.777395 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt"] Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.822985 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9vrr\" (UniqueName: \"kubernetes.io/projected/4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716-kube-api-access-f9vrr\") pod \"metallb-operator-webhook-server-57ffff8ff7-k42kt\" (UID: \"4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716\") " pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.823259 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716-apiservice-cert\") pod \"metallb-operator-webhook-server-57ffff8ff7-k42kt\" (UID: \"4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716\") " pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.823351 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716-webhook-cert\") pod \"metallb-operator-webhook-server-57ffff8ff7-k42kt\" (UID: \"4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716\") " pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.924397 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716-apiservice-cert\") pod \"metallb-operator-webhook-server-57ffff8ff7-k42kt\" (UID: \"4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716\") " pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.924771 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716-webhook-cert\") pod \"metallb-operator-webhook-server-57ffff8ff7-k42kt\" (UID: \"4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716\") " pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.924817 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9vrr\" (UniqueName: \"kubernetes.io/projected/4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716-kube-api-access-f9vrr\") pod \"metallb-operator-webhook-server-57ffff8ff7-k42kt\" (UID: \"4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716\") " pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.937492 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716-apiservice-cert\") pod \"metallb-operator-webhook-server-57ffff8ff7-k42kt\" (UID: \"4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716\") " pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.940932 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716-webhook-cert\") pod \"metallb-operator-webhook-server-57ffff8ff7-k42kt\" (UID: \"4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716\") " pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.956687 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9vrr\" (UniqueName: \"kubernetes.io/projected/4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716-kube-api-access-f9vrr\") pod \"metallb-operator-webhook-server-57ffff8ff7-k42kt\" (UID: \"4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716\") " pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:10 crc kubenswrapper[4943]: I1204 10:30:10.973606 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:11 crc kubenswrapper[4943]: I1204 10:30:11.036927 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls"] Dec 04 10:30:11 crc kubenswrapper[4943]: I1204 10:30:11.194900 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt"] Dec 04 10:30:11 crc kubenswrapper[4943]: W1204 10:30:11.199383 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4dfd2ad5_8ff8_4d1e_8976_e1cbc87d1716.slice/crio-c7e251db19015f0a2b904e13fbd1cdc002d8dfcf6a1775fe2ffcbd5196b7e090 WatchSource:0}: Error finding container c7e251db19015f0a2b904e13fbd1cdc002d8dfcf6a1775fe2ffcbd5196b7e090: Status 404 returned error can't find the container with id c7e251db19015f0a2b904e13fbd1cdc002d8dfcf6a1775fe2ffcbd5196b7e090 Dec 04 10:30:12 crc kubenswrapper[4943]: I1204 10:30:12.059253 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" event={"ID":"0ca89ccd-b04f-480d-9a4e-a96b534b3ca6","Type":"ContainerStarted","Data":"cad97c70a1cbedac2a897fcaac05269d570516d373c6bd4671a1539f72535689"} Dec 04 10:30:12 crc kubenswrapper[4943]: I1204 10:30:12.061076 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" event={"ID":"4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716","Type":"ContainerStarted","Data":"c7e251db19015f0a2b904e13fbd1cdc002d8dfcf6a1775fe2ffcbd5196b7e090"} Dec 04 10:30:17 crc kubenswrapper[4943]: I1204 10:30:17.104721 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" event={"ID":"4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716","Type":"ContainerStarted","Data":"e0c3c09dbe7ec5fd23b9e20470b482e654f2b5c6d708d1db38c2f90ddb816233"} Dec 04 10:30:17 crc kubenswrapper[4943]: I1204 10:30:17.105332 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:17 crc kubenswrapper[4943]: I1204 10:30:17.106509 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" event={"ID":"0ca89ccd-b04f-480d-9a4e-a96b534b3ca6","Type":"ContainerStarted","Data":"8eaa83b40d12f3067c8e05290f3b02a6eac00e34d337a18a135f079ed4586187"} Dec 04 10:30:17 crc kubenswrapper[4943]: I1204 10:30:17.106712 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:17 crc kubenswrapper[4943]: I1204 10:30:17.124800 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" podStartSLOduration=1.962277848 podStartE2EDuration="7.124772537s" podCreationTimestamp="2025-12-04 10:30:10 +0000 UTC" firstStartedPulling="2025-12-04 10:30:11.202894119 +0000 UTC m=+859.791869987" lastFinishedPulling="2025-12-04 10:30:16.365388808 +0000 UTC m=+864.954364676" observedRunningTime="2025-12-04 10:30:17.123356679 +0000 UTC m=+865.712332557" watchObservedRunningTime="2025-12-04 10:30:17.124772537 +0000 UTC m=+865.713748405" Dec 04 10:30:17 crc kubenswrapper[4943]: I1204 10:30:17.144399 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" podStartSLOduration=1.859112911 podStartE2EDuration="7.14438329s" podCreationTimestamp="2025-12-04 10:30:10 +0000 UTC" firstStartedPulling="2025-12-04 10:30:11.061226625 +0000 UTC m=+859.650202493" lastFinishedPulling="2025-12-04 10:30:16.346497014 +0000 UTC m=+864.935472872" observedRunningTime="2025-12-04 10:30:17.143147697 +0000 UTC m=+865.732123565" watchObservedRunningTime="2025-12-04 10:30:17.14438329 +0000 UTC m=+865.733359158" Dec 04 10:30:30 crc kubenswrapper[4943]: I1204 10:30:30.979211 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-57ffff8ff7-k42kt" Dec 04 10:30:49 crc kubenswrapper[4943]: I1204 10:30:49.752064 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t2qqp"] Dec 04 10:30:49 crc kubenswrapper[4943]: I1204 10:30:49.756906 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:30:49 crc kubenswrapper[4943]: I1204 10:30:49.759660 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t2qqp"] Dec 04 10:30:49 crc kubenswrapper[4943]: I1204 10:30:49.799915 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b156b3d-7929-43c5-9628-55a56b252794-catalog-content\") pod \"certified-operators-t2qqp\" (UID: \"5b156b3d-7929-43c5-9628-55a56b252794\") " pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:30:49 crc kubenswrapper[4943]: I1204 10:30:49.800039 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b156b3d-7929-43c5-9628-55a56b252794-utilities\") pod \"certified-operators-t2qqp\" (UID: \"5b156b3d-7929-43c5-9628-55a56b252794\") " pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:30:49 crc kubenswrapper[4943]: I1204 10:30:49.800065 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkr66\" (UniqueName: \"kubernetes.io/projected/5b156b3d-7929-43c5-9628-55a56b252794-kube-api-access-xkr66\") pod \"certified-operators-t2qqp\" (UID: \"5b156b3d-7929-43c5-9628-55a56b252794\") " pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:30:49 crc kubenswrapper[4943]: I1204 10:30:49.901426 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b156b3d-7929-43c5-9628-55a56b252794-catalog-content\") pod \"certified-operators-t2qqp\" (UID: \"5b156b3d-7929-43c5-9628-55a56b252794\") " pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:30:49 crc kubenswrapper[4943]: I1204 10:30:49.901521 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b156b3d-7929-43c5-9628-55a56b252794-utilities\") pod \"certified-operators-t2qqp\" (UID: \"5b156b3d-7929-43c5-9628-55a56b252794\") " pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:30:49 crc kubenswrapper[4943]: I1204 10:30:49.901544 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkr66\" (UniqueName: \"kubernetes.io/projected/5b156b3d-7929-43c5-9628-55a56b252794-kube-api-access-xkr66\") pod \"certified-operators-t2qqp\" (UID: \"5b156b3d-7929-43c5-9628-55a56b252794\") " pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:30:49 crc kubenswrapper[4943]: I1204 10:30:49.902150 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b156b3d-7929-43c5-9628-55a56b252794-catalog-content\") pod \"certified-operators-t2qqp\" (UID: \"5b156b3d-7929-43c5-9628-55a56b252794\") " pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:30:49 crc kubenswrapper[4943]: I1204 10:30:49.902291 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b156b3d-7929-43c5-9628-55a56b252794-utilities\") pod \"certified-operators-t2qqp\" (UID: \"5b156b3d-7929-43c5-9628-55a56b252794\") " pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:30:49 crc kubenswrapper[4943]: I1204 10:30:49.920989 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkr66\" (UniqueName: \"kubernetes.io/projected/5b156b3d-7929-43c5-9628-55a56b252794-kube-api-access-xkr66\") pod \"certified-operators-t2qqp\" (UID: \"5b156b3d-7929-43c5-9628-55a56b252794\") " pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:30:50 crc kubenswrapper[4943]: I1204 10:30:50.092048 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:30:50 crc kubenswrapper[4943]: I1204 10:30:50.572898 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t2qqp"] Dec 04 10:30:50 crc kubenswrapper[4943]: I1204 10:30:50.714072 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5697dbbbd-r8zls" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.287935 4943 generic.go:334] "Generic (PLEG): container finished" podID="5b156b3d-7929-43c5-9628-55a56b252794" containerID="93e2e0212ff8a9d06a093bc41b47d080c656ab5afbe8d0e8db8b49143c81425d" exitCode=0 Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.287983 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2qqp" event={"ID":"5b156b3d-7929-43c5-9628-55a56b252794","Type":"ContainerDied","Data":"93e2e0212ff8a9d06a093bc41b47d080c656ab5afbe8d0e8db8b49143c81425d"} Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.288256 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2qqp" event={"ID":"5b156b3d-7929-43c5-9628-55a56b252794","Type":"ContainerStarted","Data":"16f703b3ab11f225d82d9c4266eb5aefc53ed6001f91bcaccb7b890ae448511f"} Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.490989 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-k2g9w"] Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.493612 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.496435 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.496810 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.497002 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-fgz56" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.522541 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4"] Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.523952 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.531072 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.555530 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4"] Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.565593 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-metrics-certs\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.565876 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6104b8e7-773f-4873-912f-5269972dfac9-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-t89n4\" (UID: \"6104b8e7-773f-4873-912f-5269972dfac9\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.566050 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-metrics\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.566133 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-frr-conf\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.566252 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-reloader\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.566360 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpxxq\" (UniqueName: \"kubernetes.io/projected/6104b8e7-773f-4873-912f-5269972dfac9-kube-api-access-tpxxq\") pod \"frr-k8s-webhook-server-7fcb986d4-t89n4\" (UID: \"6104b8e7-773f-4873-912f-5269972dfac9\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.566444 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-frr-startup\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.566544 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-frr-sockets\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.566760 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdtvc\" (UniqueName: \"kubernetes.io/projected/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-kube-api-access-qdtvc\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.590563 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-sd2np"] Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.591448 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-sd2np" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.594070 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-sw7l7" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.594337 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.594526 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.597906 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.597968 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-hm4b7"] Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.599292 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.601491 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.611417 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-hm4b7"] Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.667613 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpxxq\" (UniqueName: \"kubernetes.io/projected/6104b8e7-773f-4873-912f-5269972dfac9-kube-api-access-tpxxq\") pod \"frr-k8s-webhook-server-7fcb986d4-t89n4\" (UID: \"6104b8e7-773f-4873-912f-5269972dfac9\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.667680 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-frr-startup\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.667714 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-metrics-certs\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.667738 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpp5b\" (UniqueName: \"kubernetes.io/projected/71342ec6-c2cb-4277-a567-e13aabab1db3-kube-api-access-jpp5b\") pod \"controller-f8648f98b-hm4b7\" (UID: \"71342ec6-c2cb-4277-a567-e13aabab1db3\") " pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.667758 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-frr-sockets\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.667785 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/71342ec6-c2cb-4277-a567-e13aabab1db3-metrics-certs\") pod \"controller-f8648f98b-hm4b7\" (UID: \"71342ec6-c2cb-4277-a567-e13aabab1db3\") " pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.668045 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-memberlist\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.668149 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdtvc\" (UniqueName: \"kubernetes.io/projected/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-kube-api-access-qdtvc\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.668192 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/71342ec6-c2cb-4277-a567-e13aabab1db3-cert\") pod \"controller-f8648f98b-hm4b7\" (UID: \"71342ec6-c2cb-4277-a567-e13aabab1db3\") " pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.668252 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-metrics-certs\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.668288 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6104b8e7-773f-4873-912f-5269972dfac9-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-t89n4\" (UID: \"6104b8e7-773f-4873-912f-5269972dfac9\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.668322 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b473a6fb-38de-4766-b414-56f3c0496ff6-metallb-excludel2\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.668390 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-metrics\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.668405 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-frr-conf\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.668426 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-reloader\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.668450 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94hhm\" (UniqueName: \"kubernetes.io/projected/b473a6fb-38de-4766-b414-56f3c0496ff6-kube-api-access-94hhm\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.668819 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-frr-startup\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.669301 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-metrics\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.669554 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-frr-conf\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: E1204 10:30:51.669670 4943 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 04 10:30:51 crc kubenswrapper[4943]: E1204 10:30:51.669744 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-metrics-certs podName:07356b24-d6a8-4b4f-8d78-b73a19e05e7d nodeName:}" failed. No retries permitted until 2025-12-04 10:30:52.169721572 +0000 UTC m=+900.758697440 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-metrics-certs") pod "frr-k8s-k2g9w" (UID: "07356b24-d6a8-4b4f-8d78-b73a19e05e7d") : secret "frr-k8s-certs-secret" not found Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.669791 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-reloader\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.670513 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-frr-sockets\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.688029 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpxxq\" (UniqueName: \"kubernetes.io/projected/6104b8e7-773f-4873-912f-5269972dfac9-kube-api-access-tpxxq\") pod \"frr-k8s-webhook-server-7fcb986d4-t89n4\" (UID: \"6104b8e7-773f-4873-912f-5269972dfac9\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.696099 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6104b8e7-773f-4873-912f-5269972dfac9-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-t89n4\" (UID: \"6104b8e7-773f-4873-912f-5269972dfac9\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.699351 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdtvc\" (UniqueName: \"kubernetes.io/projected/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-kube-api-access-qdtvc\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.769912 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-memberlist\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.769998 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/71342ec6-c2cb-4277-a567-e13aabab1db3-cert\") pod \"controller-f8648f98b-hm4b7\" (UID: \"71342ec6-c2cb-4277-a567-e13aabab1db3\") " pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.770053 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b473a6fb-38de-4766-b414-56f3c0496ff6-metallb-excludel2\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.770091 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94hhm\" (UniqueName: \"kubernetes.io/projected/b473a6fb-38de-4766-b414-56f3c0496ff6-kube-api-access-94hhm\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.770118 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-metrics-certs\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.770133 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpp5b\" (UniqueName: \"kubernetes.io/projected/71342ec6-c2cb-4277-a567-e13aabab1db3-kube-api-access-jpp5b\") pod \"controller-f8648f98b-hm4b7\" (UID: \"71342ec6-c2cb-4277-a567-e13aabab1db3\") " pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.770151 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/71342ec6-c2cb-4277-a567-e13aabab1db3-metrics-certs\") pod \"controller-f8648f98b-hm4b7\" (UID: \"71342ec6-c2cb-4277-a567-e13aabab1db3\") " pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:30:51 crc kubenswrapper[4943]: E1204 10:30:51.770297 4943 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 04 10:30:51 crc kubenswrapper[4943]: E1204 10:30:51.770348 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/71342ec6-c2cb-4277-a567-e13aabab1db3-metrics-certs podName:71342ec6-c2cb-4277-a567-e13aabab1db3 nodeName:}" failed. No retries permitted until 2025-12-04 10:30:52.270333619 +0000 UTC m=+900.859309487 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/71342ec6-c2cb-4277-a567-e13aabab1db3-metrics-certs") pod "controller-f8648f98b-hm4b7" (UID: "71342ec6-c2cb-4277-a567-e13aabab1db3") : secret "controller-certs-secret" not found Dec 04 10:30:51 crc kubenswrapper[4943]: E1204 10:30:51.770391 4943 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 10:30:51 crc kubenswrapper[4943]: E1204 10:30:51.770470 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-memberlist podName:b473a6fb-38de-4766-b414-56f3c0496ff6 nodeName:}" failed. No retries permitted until 2025-12-04 10:30:52.270462913 +0000 UTC m=+900.859438781 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-memberlist") pod "speaker-sd2np" (UID: "b473a6fb-38de-4766-b414-56f3c0496ff6") : secret "metallb-memberlist" not found Dec 04 10:30:51 crc kubenswrapper[4943]: E1204 10:30:51.770511 4943 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 04 10:30:51 crc kubenswrapper[4943]: E1204 10:30:51.770528 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-metrics-certs podName:b473a6fb-38de-4766-b414-56f3c0496ff6 nodeName:}" failed. No retries permitted until 2025-12-04 10:30:52.270523454 +0000 UTC m=+900.859499322 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-metrics-certs") pod "speaker-sd2np" (UID: "b473a6fb-38de-4766-b414-56f3c0496ff6") : secret "speaker-certs-secret" not found Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.771736 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b473a6fb-38de-4766-b414-56f3c0496ff6-metallb-excludel2\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.773839 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/71342ec6-c2cb-4277-a567-e13aabab1db3-cert\") pod \"controller-f8648f98b-hm4b7\" (UID: \"71342ec6-c2cb-4277-a567-e13aabab1db3\") " pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.794850 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpp5b\" (UniqueName: \"kubernetes.io/projected/71342ec6-c2cb-4277-a567-e13aabab1db3-kube-api-access-jpp5b\") pod \"controller-f8648f98b-hm4b7\" (UID: \"71342ec6-c2cb-4277-a567-e13aabab1db3\") " pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.803029 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94hhm\" (UniqueName: \"kubernetes.io/projected/b473a6fb-38de-4766-b414-56f3c0496ff6-kube-api-access-94hhm\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:51 crc kubenswrapper[4943]: I1204 10:30:51.860241 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.046088 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4"] Dec 04 10:30:52 crc kubenswrapper[4943]: W1204 10:30:52.051786 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6104b8e7_773f_4873_912f_5269972dfac9.slice/crio-ab962a8084acda3e3892caeb474485b3cd862922a0e54874994ea9c499d570a7 WatchSource:0}: Error finding container ab962a8084acda3e3892caeb474485b3cd862922a0e54874994ea9c499d570a7: Status 404 returned error can't find the container with id ab962a8084acda3e3892caeb474485b3cd862922a0e54874994ea9c499d570a7 Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.175679 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-metrics-certs\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.180766 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/07356b24-d6a8-4b4f-8d78-b73a19e05e7d-metrics-certs\") pod \"frr-k8s-k2g9w\" (UID: \"07356b24-d6a8-4b4f-8d78-b73a19e05e7d\") " pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.277153 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-metrics-certs\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.277239 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/71342ec6-c2cb-4277-a567-e13aabab1db3-metrics-certs\") pod \"controller-f8648f98b-hm4b7\" (UID: \"71342ec6-c2cb-4277-a567-e13aabab1db3\") " pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.277273 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-memberlist\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:52 crc kubenswrapper[4943]: E1204 10:30:52.277471 4943 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 10:30:52 crc kubenswrapper[4943]: E1204 10:30:52.277568 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-memberlist podName:b473a6fb-38de-4766-b414-56f3c0496ff6 nodeName:}" failed. No retries permitted until 2025-12-04 10:30:53.277519876 +0000 UTC m=+901.866495744 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-memberlist") pod "speaker-sd2np" (UID: "b473a6fb-38de-4766-b414-56f3c0496ff6") : secret "metallb-memberlist" not found Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.280030 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-metrics-certs\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.281742 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/71342ec6-c2cb-4277-a567-e13aabab1db3-metrics-certs\") pod \"controller-f8648f98b-hm4b7\" (UID: \"71342ec6-c2cb-4277-a567-e13aabab1db3\") " pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.296045 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" event={"ID":"6104b8e7-773f-4873-912f-5269972dfac9","Type":"ContainerStarted","Data":"ab962a8084acda3e3892caeb474485b3cd862922a0e54874994ea9c499d570a7"} Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.301257 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2qqp" event={"ID":"5b156b3d-7929-43c5-9628-55a56b252794","Type":"ContainerStarted","Data":"06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca"} Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.421182 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.522011 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:30:52 crc kubenswrapper[4943]: I1204 10:30:52.793920 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-hm4b7"] Dec 04 10:30:53 crc kubenswrapper[4943]: E1204 10:30:53.286433 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b156b3d_7929_43c5_9628_55a56b252794.slice/crio-06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b156b3d_7929_43c5_9628_55a56b252794.slice/crio-conmon-06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca.scope\": RecentStats: unable to find data in memory cache]" Dec 04 10:30:53 crc kubenswrapper[4943]: I1204 10:30:53.306422 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k2g9w" event={"ID":"07356b24-d6a8-4b4f-8d78-b73a19e05e7d","Type":"ContainerStarted","Data":"64e363da7ec97ccb74f2d40e0d9b31e402d91f9749253cadd225f283c84020ec"} Dec 04 10:30:53 crc kubenswrapper[4943]: I1204 10:30:53.306544 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-memberlist\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:53 crc kubenswrapper[4943]: I1204 10:30:53.307639 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-hm4b7" event={"ID":"71342ec6-c2cb-4277-a567-e13aabab1db3","Type":"ContainerStarted","Data":"942f062877256a193b7454c69a605ccc7691fa22abb7ded17db2fb361ac0bdfc"} Dec 04 10:30:53 crc kubenswrapper[4943]: I1204 10:30:53.307664 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-hm4b7" event={"ID":"71342ec6-c2cb-4277-a567-e13aabab1db3","Type":"ContainerStarted","Data":"ba3b8ec9c74fe35f047ed1ab06264b787225dd0b95685ed24ee25307034d3ac5"} Dec 04 10:30:53 crc kubenswrapper[4943]: I1204 10:30:53.309437 4943 generic.go:334] "Generic (PLEG): container finished" podID="5b156b3d-7929-43c5-9628-55a56b252794" containerID="06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca" exitCode=0 Dec 04 10:30:53 crc kubenswrapper[4943]: I1204 10:30:53.309485 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2qqp" event={"ID":"5b156b3d-7929-43c5-9628-55a56b252794","Type":"ContainerDied","Data":"06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca"} Dec 04 10:30:53 crc kubenswrapper[4943]: I1204 10:30:53.315917 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b473a6fb-38de-4766-b414-56f3c0496ff6-memberlist\") pod \"speaker-sd2np\" (UID: \"b473a6fb-38de-4766-b414-56f3c0496ff6\") " pod="metallb-system/speaker-sd2np" Dec 04 10:30:53 crc kubenswrapper[4943]: I1204 10:30:53.413258 4943 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-sw7l7" Dec 04 10:30:53 crc kubenswrapper[4943]: I1204 10:30:53.421991 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-sd2np" Dec 04 10:30:54 crc kubenswrapper[4943]: I1204 10:30:54.316211 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sd2np" event={"ID":"b473a6fb-38de-4766-b414-56f3c0496ff6","Type":"ContainerStarted","Data":"bd26dd626f44c04dee736d03f84ef08123a1c92af00a731b359624a13ef64fe4"} Dec 04 10:30:54 crc kubenswrapper[4943]: I1204 10:30:54.316508 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sd2np" event={"ID":"b473a6fb-38de-4766-b414-56f3c0496ff6","Type":"ContainerStarted","Data":"5204618befa7af7f695caa23991ee18099dd5c61699cc9cf712c016acc50e4b7"} Dec 04 10:30:55 crc kubenswrapper[4943]: I1204 10:30:55.326239 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2qqp" event={"ID":"5b156b3d-7929-43c5-9628-55a56b252794","Type":"ContainerStarted","Data":"4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8"} Dec 04 10:30:55 crc kubenswrapper[4943]: I1204 10:30:55.350707 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t2qqp" podStartSLOduration=3.287706624 podStartE2EDuration="6.350687438s" podCreationTimestamp="2025-12-04 10:30:49 +0000 UTC" firstStartedPulling="2025-12-04 10:30:51.289651932 +0000 UTC m=+899.878627800" lastFinishedPulling="2025-12-04 10:30:54.352632746 +0000 UTC m=+902.941608614" observedRunningTime="2025-12-04 10:30:55.349401803 +0000 UTC m=+903.938377671" watchObservedRunningTime="2025-12-04 10:30:55.350687438 +0000 UTC m=+903.939663316" Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.093366 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.093951 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.136275 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.354072 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sd2np" event={"ID":"b473a6fb-38de-4766-b414-56f3c0496ff6","Type":"ContainerStarted","Data":"f3330d120b7c5b9951dd01cedef5af97420abfdecf76eb3a35a3baeb75dd0b7e"} Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.354208 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-sd2np" Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.355721 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" event={"ID":"6104b8e7-773f-4873-912f-5269972dfac9","Type":"ContainerStarted","Data":"a86028a79441c49da248fa77bbb89e4344c83945879f90ba0e93222ef2d657e6"} Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.357489 4943 generic.go:334] "Generic (PLEG): container finished" podID="07356b24-d6a8-4b4f-8d78-b73a19e05e7d" containerID="e15043dcafda40fc518891a2951e219bd567299b89bab5d8be98f1f8b1c2b765" exitCode=0 Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.357578 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k2g9w" event={"ID":"07356b24-d6a8-4b4f-8d78-b73a19e05e7d","Type":"ContainerDied","Data":"e15043dcafda40fc518891a2951e219bd567299b89bab5d8be98f1f8b1c2b765"} Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.359585 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-hm4b7" event={"ID":"71342ec6-c2cb-4277-a567-e13aabab1db3","Type":"ContainerStarted","Data":"b528ae3282fee0b66e7f81b5198023f554e462fe07c75db7cfc87d2f5db39949"} Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.383562 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-sd2np" podStartSLOduration=3.804091396 podStartE2EDuration="9.383513851s" podCreationTimestamp="2025-12-04 10:30:51 +0000 UTC" firstStartedPulling="2025-12-04 10:30:53.720174422 +0000 UTC m=+902.309150290" lastFinishedPulling="2025-12-04 10:30:59.299596877 +0000 UTC m=+907.888572745" observedRunningTime="2025-12-04 10:31:00.382574974 +0000 UTC m=+908.971551112" watchObservedRunningTime="2025-12-04 10:31:00.383513851 +0000 UTC m=+908.972489719" Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.404475 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-hm4b7" podStartSLOduration=3.031168064 podStartE2EDuration="9.404452765s" podCreationTimestamp="2025-12-04 10:30:51 +0000 UTC" firstStartedPulling="2025-12-04 10:30:52.926299195 +0000 UTC m=+901.515275063" lastFinishedPulling="2025-12-04 10:30:59.299583906 +0000 UTC m=+907.888559764" observedRunningTime="2025-12-04 10:31:00.402402497 +0000 UTC m=+908.991378365" watchObservedRunningTime="2025-12-04 10:31:00.404452765 +0000 UTC m=+908.993428643" Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.405755 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.444542 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" podStartSLOduration=2.198489629 podStartE2EDuration="9.444520502s" podCreationTimestamp="2025-12-04 10:30:51 +0000 UTC" firstStartedPulling="2025-12-04 10:30:52.053578504 +0000 UTC m=+900.642554372" lastFinishedPulling="2025-12-04 10:30:59.299609367 +0000 UTC m=+907.888585245" observedRunningTime="2025-12-04 10:31:00.440371704 +0000 UTC m=+909.029347582" watchObservedRunningTime="2025-12-04 10:31:00.444520502 +0000 UTC m=+909.033496370" Dec 04 10:31:00 crc kubenswrapper[4943]: I1204 10:31:00.491355 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t2qqp"] Dec 04 10:31:01 crc kubenswrapper[4943]: I1204 10:31:01.366399 4943 generic.go:334] "Generic (PLEG): container finished" podID="07356b24-d6a8-4b4f-8d78-b73a19e05e7d" containerID="8b0af41df6d92a3feee5a7e9bb259fca86d2eb099bc95624d2ce655b0cf3562a" exitCode=0 Dec 04 10:31:01 crc kubenswrapper[4943]: I1204 10:31:01.366501 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k2g9w" event={"ID":"07356b24-d6a8-4b4f-8d78-b73a19e05e7d","Type":"ContainerDied","Data":"8b0af41df6d92a3feee5a7e9bb259fca86d2eb099bc95624d2ce655b0cf3562a"} Dec 04 10:31:01 crc kubenswrapper[4943]: I1204 10:31:01.367469 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:31:01 crc kubenswrapper[4943]: I1204 10:31:01.367494 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" Dec 04 10:31:02 crc kubenswrapper[4943]: I1204 10:31:02.373934 4943 generic.go:334] "Generic (PLEG): container finished" podID="07356b24-d6a8-4b4f-8d78-b73a19e05e7d" containerID="25c8be62cc73f93dd465a21a322af1f51e8d4c029e6b9c74866ed1f2b258da60" exitCode=0 Dec 04 10:31:02 crc kubenswrapper[4943]: I1204 10:31:02.374023 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k2g9w" event={"ID":"07356b24-d6a8-4b4f-8d78-b73a19e05e7d","Type":"ContainerDied","Data":"25c8be62cc73f93dd465a21a322af1f51e8d4c029e6b9c74866ed1f2b258da60"} Dec 04 10:31:02 crc kubenswrapper[4943]: I1204 10:31:02.375276 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t2qqp" podUID="5b156b3d-7929-43c5-9628-55a56b252794" containerName="registry-server" containerID="cri-o://4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8" gracePeriod=2 Dec 04 10:31:02 crc kubenswrapper[4943]: I1204 10:31:02.377782 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-hm4b7" Dec 04 10:31:02 crc kubenswrapper[4943]: I1204 10:31:02.761758 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:31:02 crc kubenswrapper[4943]: I1204 10:31:02.955064 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b156b3d-7929-43c5-9628-55a56b252794-catalog-content\") pod \"5b156b3d-7929-43c5-9628-55a56b252794\" (UID: \"5b156b3d-7929-43c5-9628-55a56b252794\") " Dec 04 10:31:02 crc kubenswrapper[4943]: I1204 10:31:02.956103 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b156b3d-7929-43c5-9628-55a56b252794-utilities\") pod \"5b156b3d-7929-43c5-9628-55a56b252794\" (UID: \"5b156b3d-7929-43c5-9628-55a56b252794\") " Dec 04 10:31:02 crc kubenswrapper[4943]: I1204 10:31:02.956173 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkr66\" (UniqueName: \"kubernetes.io/projected/5b156b3d-7929-43c5-9628-55a56b252794-kube-api-access-xkr66\") pod \"5b156b3d-7929-43c5-9628-55a56b252794\" (UID: \"5b156b3d-7929-43c5-9628-55a56b252794\") " Dec 04 10:31:02 crc kubenswrapper[4943]: I1204 10:31:02.957244 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b156b3d-7929-43c5-9628-55a56b252794-utilities" (OuterVolumeSpecName: "utilities") pod "5b156b3d-7929-43c5-9628-55a56b252794" (UID: "5b156b3d-7929-43c5-9628-55a56b252794"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:31:02 crc kubenswrapper[4943]: I1204 10:31:02.965619 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b156b3d-7929-43c5-9628-55a56b252794-kube-api-access-xkr66" (OuterVolumeSpecName: "kube-api-access-xkr66") pod "5b156b3d-7929-43c5-9628-55a56b252794" (UID: "5b156b3d-7929-43c5-9628-55a56b252794"). InnerVolumeSpecName "kube-api-access-xkr66". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.057546 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b156b3d-7929-43c5-9628-55a56b252794-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.057590 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkr66\" (UniqueName: \"kubernetes.io/projected/5b156b3d-7929-43c5-9628-55a56b252794-kube-api-access-xkr66\") on node \"crc\" DevicePath \"\"" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.157798 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b156b3d-7929-43c5-9628-55a56b252794-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b156b3d-7929-43c5-9628-55a56b252794" (UID: "5b156b3d-7929-43c5-9628-55a56b252794"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.158570 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b156b3d-7929-43c5-9628-55a56b252794-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.382641 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k2g9w" event={"ID":"07356b24-d6a8-4b4f-8d78-b73a19e05e7d","Type":"ContainerStarted","Data":"b65b1bfd418d383ae2caea13a854408423a60bb7f4ab41add4aba08bf93fcfe0"} Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.382690 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k2g9w" event={"ID":"07356b24-d6a8-4b4f-8d78-b73a19e05e7d","Type":"ContainerStarted","Data":"d4b3f900bdc2220a99cfcb674d6ca21be93d414aac05eb527bfbfad173f1c8f9"} Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.382706 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k2g9w" event={"ID":"07356b24-d6a8-4b4f-8d78-b73a19e05e7d","Type":"ContainerStarted","Data":"3988d36aa0207aae780732b069b09e40c4724c1951a6447df151b48ce0c83943"} Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.384785 4943 generic.go:334] "Generic (PLEG): container finished" podID="5b156b3d-7929-43c5-9628-55a56b252794" containerID="4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8" exitCode=0 Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.384835 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2qqp" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.384852 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2qqp" event={"ID":"5b156b3d-7929-43c5-9628-55a56b252794","Type":"ContainerDied","Data":"4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8"} Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.384897 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2qqp" event={"ID":"5b156b3d-7929-43c5-9628-55a56b252794","Type":"ContainerDied","Data":"16f703b3ab11f225d82d9c4266eb5aefc53ed6001f91bcaccb7b890ae448511f"} Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.384919 4943 scope.go:117] "RemoveContainer" containerID="4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.409441 4943 scope.go:117] "RemoveContainer" containerID="06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.415376 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t2qqp"] Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.422390 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t2qqp"] Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.439307 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-sd2np" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.439385 4943 scope.go:117] "RemoveContainer" containerID="93e2e0212ff8a9d06a093bc41b47d080c656ab5afbe8d0e8db8b49143c81425d" Dec 04 10:31:03 crc kubenswrapper[4943]: E1204 10:31:03.456947 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b156b3d_7929_43c5_9628_55a56b252794.slice/crio-conmon-06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b156b3d_7929_43c5_9628_55a56b252794.slice/crio-06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca.scope\": RecentStats: unable to find data in memory cache]" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.486667 4943 scope.go:117] "RemoveContainer" containerID="4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8" Dec 04 10:31:03 crc kubenswrapper[4943]: E1204 10:31:03.487049 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8\": container with ID starting with 4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8 not found: ID does not exist" containerID="4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.487108 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8"} err="failed to get container status \"4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8\": rpc error: code = NotFound desc = could not find container \"4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8\": container with ID starting with 4990fc269fbdf7c8c13cb00fd89e44fb5471349c37742d9da78a9f8a8c845cc8 not found: ID does not exist" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.487134 4943 scope.go:117] "RemoveContainer" containerID="06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca" Dec 04 10:31:03 crc kubenswrapper[4943]: E1204 10:31:03.487370 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca\": container with ID starting with 06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca not found: ID does not exist" containerID="06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.487394 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca"} err="failed to get container status \"06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca\": rpc error: code = NotFound desc = could not find container \"06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca\": container with ID starting with 06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca not found: ID does not exist" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.487407 4943 scope.go:117] "RemoveContainer" containerID="93e2e0212ff8a9d06a093bc41b47d080c656ab5afbe8d0e8db8b49143c81425d" Dec 04 10:31:03 crc kubenswrapper[4943]: E1204 10:31:03.487911 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93e2e0212ff8a9d06a093bc41b47d080c656ab5afbe8d0e8db8b49143c81425d\": container with ID starting with 93e2e0212ff8a9d06a093bc41b47d080c656ab5afbe8d0e8db8b49143c81425d not found: ID does not exist" containerID="93e2e0212ff8a9d06a093bc41b47d080c656ab5afbe8d0e8db8b49143c81425d" Dec 04 10:31:03 crc kubenswrapper[4943]: I1204 10:31:03.487969 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e2e0212ff8a9d06a093bc41b47d080c656ab5afbe8d0e8db8b49143c81425d"} err="failed to get container status \"93e2e0212ff8a9d06a093bc41b47d080c656ab5afbe8d0e8db8b49143c81425d\": rpc error: code = NotFound desc = could not find container \"93e2e0212ff8a9d06a093bc41b47d080c656ab5afbe8d0e8db8b49143c81425d\": container with ID starting with 93e2e0212ff8a9d06a093bc41b47d080c656ab5afbe8d0e8db8b49143c81425d not found: ID does not exist" Dec 04 10:31:04 crc kubenswrapper[4943]: I1204 10:31:04.395248 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k2g9w" event={"ID":"07356b24-d6a8-4b4f-8d78-b73a19e05e7d","Type":"ContainerStarted","Data":"e9ee892504cc74b2b98dc96eb40d3c68627d229be4880f48e1ba21e43b9613dc"} Dec 04 10:31:04 crc kubenswrapper[4943]: I1204 10:31:04.395594 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k2g9w" event={"ID":"07356b24-d6a8-4b4f-8d78-b73a19e05e7d","Type":"ContainerStarted","Data":"e71d5e253d3fd43431ebbf1f55f9ce89936b90a27686aef6e69ec7b93fef5e42"} Dec 04 10:31:04 crc kubenswrapper[4943]: I1204 10:31:04.395615 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-k2g9w" event={"ID":"07356b24-d6a8-4b4f-8d78-b73a19e05e7d","Type":"ContainerStarted","Data":"eaf8739cb3d602591cac7670c51f21d7533a71b0b98ac184bc9ea8edd891564a"} Dec 04 10:31:04 crc kubenswrapper[4943]: I1204 10:31:04.420830 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-k2g9w" podStartSLOduration=6.658813273 podStartE2EDuration="13.420804191s" podCreationTimestamp="2025-12-04 10:30:51 +0000 UTC" firstStartedPulling="2025-12-04 10:30:52.574010021 +0000 UTC m=+901.162985889" lastFinishedPulling="2025-12-04 10:30:59.336000939 +0000 UTC m=+907.924976807" observedRunningTime="2025-12-04 10:31:04.417697513 +0000 UTC m=+913.006673401" watchObservedRunningTime="2025-12-04 10:31:04.420804191 +0000 UTC m=+913.009780059" Dec 04 10:31:04 crc kubenswrapper[4943]: I1204 10:31:04.574086 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b156b3d-7929-43c5-9628-55a56b252794" path="/var/lib/kubelet/pods/5b156b3d-7929-43c5-9628-55a56b252794/volumes" Dec 04 10:31:05 crc kubenswrapper[4943]: I1204 10:31:05.401721 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:31:07 crc kubenswrapper[4943]: I1204 10:31:07.422256 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:31:07 crc kubenswrapper[4943]: I1204 10:31:07.457033 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.045244 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xvmdn"] Dec 04 10:31:08 crc kubenswrapper[4943]: E1204 10:31:08.045738 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b156b3d-7929-43c5-9628-55a56b252794" containerName="extract-content" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.045755 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b156b3d-7929-43c5-9628-55a56b252794" containerName="extract-content" Dec 04 10:31:08 crc kubenswrapper[4943]: E1204 10:31:08.045782 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b156b3d-7929-43c5-9628-55a56b252794" containerName="extract-utilities" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.045796 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b156b3d-7929-43c5-9628-55a56b252794" containerName="extract-utilities" Dec 04 10:31:08 crc kubenswrapper[4943]: E1204 10:31:08.045821 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b156b3d-7929-43c5-9628-55a56b252794" containerName="registry-server" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.045829 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b156b3d-7929-43c5-9628-55a56b252794" containerName="registry-server" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.046089 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b156b3d-7929-43c5-9628-55a56b252794" containerName="registry-server" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.056457 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.064363 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvmdn"] Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.128175 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3602eccf-ee26-4836-a73b-15f8e65424b3-catalog-content\") pod \"community-operators-xvmdn\" (UID: \"3602eccf-ee26-4836-a73b-15f8e65424b3\") " pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.129262 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdvlj\" (UniqueName: \"kubernetes.io/projected/3602eccf-ee26-4836-a73b-15f8e65424b3-kube-api-access-zdvlj\") pod \"community-operators-xvmdn\" (UID: \"3602eccf-ee26-4836-a73b-15f8e65424b3\") " pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.129456 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3602eccf-ee26-4836-a73b-15f8e65424b3-utilities\") pod \"community-operators-xvmdn\" (UID: \"3602eccf-ee26-4836-a73b-15f8e65424b3\") " pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.230889 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3602eccf-ee26-4836-a73b-15f8e65424b3-catalog-content\") pod \"community-operators-xvmdn\" (UID: \"3602eccf-ee26-4836-a73b-15f8e65424b3\") " pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.230977 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdvlj\" (UniqueName: \"kubernetes.io/projected/3602eccf-ee26-4836-a73b-15f8e65424b3-kube-api-access-zdvlj\") pod \"community-operators-xvmdn\" (UID: \"3602eccf-ee26-4836-a73b-15f8e65424b3\") " pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.230997 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3602eccf-ee26-4836-a73b-15f8e65424b3-utilities\") pod \"community-operators-xvmdn\" (UID: \"3602eccf-ee26-4836-a73b-15f8e65424b3\") " pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.231535 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3602eccf-ee26-4836-a73b-15f8e65424b3-catalog-content\") pod \"community-operators-xvmdn\" (UID: \"3602eccf-ee26-4836-a73b-15f8e65424b3\") " pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.231565 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3602eccf-ee26-4836-a73b-15f8e65424b3-utilities\") pod \"community-operators-xvmdn\" (UID: \"3602eccf-ee26-4836-a73b-15f8e65424b3\") " pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.252316 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdvlj\" (UniqueName: \"kubernetes.io/projected/3602eccf-ee26-4836-a73b-15f8e65424b3-kube-api-access-zdvlj\") pod \"community-operators-xvmdn\" (UID: \"3602eccf-ee26-4836-a73b-15f8e65424b3\") " pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.385156 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:08 crc kubenswrapper[4943]: I1204 10:31:08.638970 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xvmdn"] Dec 04 10:31:08 crc kubenswrapper[4943]: W1204 10:31:08.647374 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3602eccf_ee26_4836_a73b_15f8e65424b3.slice/crio-4d8c0056a9370ce8f50ba20b487a47ad854460fc9d5484c3ca781a0f77fbc43d WatchSource:0}: Error finding container 4d8c0056a9370ce8f50ba20b487a47ad854460fc9d5484c3ca781a0f77fbc43d: Status 404 returned error can't find the container with id 4d8c0056a9370ce8f50ba20b487a47ad854460fc9d5484c3ca781a0f77fbc43d Dec 04 10:31:09 crc kubenswrapper[4943]: I1204 10:31:09.424036 4943 generic.go:334] "Generic (PLEG): container finished" podID="3602eccf-ee26-4836-a73b-15f8e65424b3" containerID="342d4a2523389e7ea3aee0394afe1fc92d8c23e674a7cecbba50864d7c92b9f4" exitCode=0 Dec 04 10:31:09 crc kubenswrapper[4943]: I1204 10:31:09.424143 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvmdn" event={"ID":"3602eccf-ee26-4836-a73b-15f8e65424b3","Type":"ContainerDied","Data":"342d4a2523389e7ea3aee0394afe1fc92d8c23e674a7cecbba50864d7c92b9f4"} Dec 04 10:31:09 crc kubenswrapper[4943]: I1204 10:31:09.424382 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvmdn" event={"ID":"3602eccf-ee26-4836-a73b-15f8e65424b3","Type":"ContainerStarted","Data":"4d8c0056a9370ce8f50ba20b487a47ad854460fc9d5484c3ca781a0f77fbc43d"} Dec 04 10:31:10 crc kubenswrapper[4943]: I1204 10:31:10.434495 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvmdn" event={"ID":"3602eccf-ee26-4836-a73b-15f8e65424b3","Type":"ContainerStarted","Data":"f54355e898831ce8d1b46f69563bb7d2dc0b129752fd423bd0c2622d3928fc7f"} Dec 04 10:31:11 crc kubenswrapper[4943]: I1204 10:31:11.441719 4943 generic.go:334] "Generic (PLEG): container finished" podID="3602eccf-ee26-4836-a73b-15f8e65424b3" containerID="f54355e898831ce8d1b46f69563bb7d2dc0b129752fd423bd0c2622d3928fc7f" exitCode=0 Dec 04 10:31:11 crc kubenswrapper[4943]: I1204 10:31:11.441769 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvmdn" event={"ID":"3602eccf-ee26-4836-a73b-15f8e65424b3","Type":"ContainerDied","Data":"f54355e898831ce8d1b46f69563bb7d2dc0b129752fd423bd0c2622d3928fc7f"} Dec 04 10:31:11 crc kubenswrapper[4943]: I1204 10:31:11.863842 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-t89n4" Dec 04 10:31:12 crc kubenswrapper[4943]: I1204 10:31:12.444550 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-k2g9w" Dec 04 10:31:12 crc kubenswrapper[4943]: I1204 10:31:12.453857 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvmdn" event={"ID":"3602eccf-ee26-4836-a73b-15f8e65424b3","Type":"ContainerStarted","Data":"c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd"} Dec 04 10:31:12 crc kubenswrapper[4943]: I1204 10:31:12.491209 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xvmdn" podStartSLOduration=1.980076855 podStartE2EDuration="4.491162332s" podCreationTimestamp="2025-12-04 10:31:08 +0000 UTC" firstStartedPulling="2025-12-04 10:31:09.426379845 +0000 UTC m=+918.015355713" lastFinishedPulling="2025-12-04 10:31:11.937465312 +0000 UTC m=+920.526441190" observedRunningTime="2025-12-04 10:31:12.488368913 +0000 UTC m=+921.077344801" watchObservedRunningTime="2025-12-04 10:31:12.491162332 +0000 UTC m=+921.080138200" Dec 04 10:31:13 crc kubenswrapper[4943]: I1204 10:31:13.221156 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-kf8zp"] Dec 04 10:31:13 crc kubenswrapper[4943]: I1204 10:31:13.222096 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-kf8zp" Dec 04 10:31:13 crc kubenswrapper[4943]: I1204 10:31:13.224190 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 04 10:31:13 crc kubenswrapper[4943]: I1204 10:31:13.224357 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-vfntr" Dec 04 10:31:13 crc kubenswrapper[4943]: I1204 10:31:13.224693 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 04 10:31:13 crc kubenswrapper[4943]: I1204 10:31:13.254841 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-kf8zp"] Dec 04 10:31:13 crc kubenswrapper[4943]: I1204 10:31:13.306125 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bdkt\" (UniqueName: \"kubernetes.io/projected/46cb23ec-4521-4cff-aae6-3183a0c6b5bb-kube-api-access-4bdkt\") pod \"mariadb-operator-index-kf8zp\" (UID: \"46cb23ec-4521-4cff-aae6-3183a0c6b5bb\") " pod="openstack-operators/mariadb-operator-index-kf8zp" Dec 04 10:31:13 crc kubenswrapper[4943]: I1204 10:31:13.406862 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bdkt\" (UniqueName: \"kubernetes.io/projected/46cb23ec-4521-4cff-aae6-3183a0c6b5bb-kube-api-access-4bdkt\") pod \"mariadb-operator-index-kf8zp\" (UID: \"46cb23ec-4521-4cff-aae6-3183a0c6b5bb\") " pod="openstack-operators/mariadb-operator-index-kf8zp" Dec 04 10:31:13 crc kubenswrapper[4943]: I1204 10:31:13.423664 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bdkt\" (UniqueName: \"kubernetes.io/projected/46cb23ec-4521-4cff-aae6-3183a0c6b5bb-kube-api-access-4bdkt\") pod \"mariadb-operator-index-kf8zp\" (UID: \"46cb23ec-4521-4cff-aae6-3183a0c6b5bb\") " pod="openstack-operators/mariadb-operator-index-kf8zp" Dec 04 10:31:13 crc kubenswrapper[4943]: I1204 10:31:13.551038 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-kf8zp" Dec 04 10:31:13 crc kubenswrapper[4943]: E1204 10:31:13.581172 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b156b3d_7929_43c5_9628_55a56b252794.slice/crio-06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b156b3d_7929_43c5_9628_55a56b252794.slice/crio-conmon-06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca.scope\": RecentStats: unable to find data in memory cache]" Dec 04 10:31:13 crc kubenswrapper[4943]: I1204 10:31:13.986373 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-kf8zp"] Dec 04 10:31:14 crc kubenswrapper[4943]: I1204 10:31:14.468060 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-kf8zp" event={"ID":"46cb23ec-4521-4cff-aae6-3183a0c6b5bb","Type":"ContainerStarted","Data":"33f525a815f68e5018394e1aeecec0704662043bd3c2ddacb3a7fb7eaa3b3b40"} Dec 04 10:31:15 crc kubenswrapper[4943]: I1204 10:31:15.476136 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-kf8zp" event={"ID":"46cb23ec-4521-4cff-aae6-3183a0c6b5bb","Type":"ContainerStarted","Data":"19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508"} Dec 04 10:31:15 crc kubenswrapper[4943]: I1204 10:31:15.491989 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-kf8zp" podStartSLOduration=1.6649717590000002 podStartE2EDuration="2.491948153s" podCreationTimestamp="2025-12-04 10:31:13 +0000 UTC" firstStartedPulling="2025-12-04 10:31:14.006874237 +0000 UTC m=+922.595850105" lastFinishedPulling="2025-12-04 10:31:14.833850631 +0000 UTC m=+923.422826499" observedRunningTime="2025-12-04 10:31:15.490452521 +0000 UTC m=+924.079428389" watchObservedRunningTime="2025-12-04 10:31:15.491948153 +0000 UTC m=+924.080924031" Dec 04 10:31:18 crc kubenswrapper[4943]: I1204 10:31:18.386128 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:18 crc kubenswrapper[4943]: I1204 10:31:18.386493 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:18 crc kubenswrapper[4943]: I1204 10:31:18.410723 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-kf8zp"] Dec 04 10:31:18 crc kubenswrapper[4943]: I1204 10:31:18.411065 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-kf8zp" podUID="46cb23ec-4521-4cff-aae6-3183a0c6b5bb" containerName="registry-server" containerID="cri-o://19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508" gracePeriod=2 Dec 04 10:31:18 crc kubenswrapper[4943]: I1204 10:31:18.434124 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:18 crc kubenswrapper[4943]: I1204 10:31:18.528068 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.014193 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-q9s4x"] Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.015385 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-q9s4x" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.021528 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-q9s4x"] Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.118175 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbdh8\" (UniqueName: \"kubernetes.io/projected/c076efe8-6293-4cd9-a27d-d6196e44b0e2-kube-api-access-gbdh8\") pod \"mariadb-operator-index-q9s4x\" (UID: \"c076efe8-6293-4cd9-a27d-d6196e44b0e2\") " pod="openstack-operators/mariadb-operator-index-q9s4x" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.219286 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbdh8\" (UniqueName: \"kubernetes.io/projected/c076efe8-6293-4cd9-a27d-d6196e44b0e2-kube-api-access-gbdh8\") pod \"mariadb-operator-index-q9s4x\" (UID: \"c076efe8-6293-4cd9-a27d-d6196e44b0e2\") " pod="openstack-operators/mariadb-operator-index-q9s4x" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.246863 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbdh8\" (UniqueName: \"kubernetes.io/projected/c076efe8-6293-4cd9-a27d-d6196e44b0e2-kube-api-access-gbdh8\") pod \"mariadb-operator-index-q9s4x\" (UID: \"c076efe8-6293-4cd9-a27d-d6196e44b0e2\") " pod="openstack-operators/mariadb-operator-index-q9s4x" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.296638 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-kf8zp" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.320311 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bdkt\" (UniqueName: \"kubernetes.io/projected/46cb23ec-4521-4cff-aae6-3183a0c6b5bb-kube-api-access-4bdkt\") pod \"46cb23ec-4521-4cff-aae6-3183a0c6b5bb\" (UID: \"46cb23ec-4521-4cff-aae6-3183a0c6b5bb\") " Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.325563 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46cb23ec-4521-4cff-aae6-3183a0c6b5bb-kube-api-access-4bdkt" (OuterVolumeSpecName: "kube-api-access-4bdkt") pod "46cb23ec-4521-4cff-aae6-3183a0c6b5bb" (UID: "46cb23ec-4521-4cff-aae6-3183a0c6b5bb"). InnerVolumeSpecName "kube-api-access-4bdkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.337060 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-q9s4x" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.416351 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xvmdn"] Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.429307 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bdkt\" (UniqueName: \"kubernetes.io/projected/46cb23ec-4521-4cff-aae6-3183a0c6b5bb-kube-api-access-4bdkt\") on node \"crc\" DevicePath \"\"" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.507686 4943 generic.go:334] "Generic (PLEG): container finished" podID="46cb23ec-4521-4cff-aae6-3183a0c6b5bb" containerID="19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508" exitCode=0 Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.509769 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-kf8zp" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.509758 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-kf8zp" event={"ID":"46cb23ec-4521-4cff-aae6-3183a0c6b5bb","Type":"ContainerDied","Data":"19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508"} Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.522378 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-kf8zp" event={"ID":"46cb23ec-4521-4cff-aae6-3183a0c6b5bb","Type":"ContainerDied","Data":"33f525a815f68e5018394e1aeecec0704662043bd3c2ddacb3a7fb7eaa3b3b40"} Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.522449 4943 scope.go:117] "RemoveContainer" containerID="19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.550962 4943 scope.go:117] "RemoveContainer" containerID="19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508" Dec 04 10:31:19 crc kubenswrapper[4943]: E1204 10:31:19.551927 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508\": container with ID starting with 19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508 not found: ID does not exist" containerID="19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.552003 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508"} err="failed to get container status \"19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508\": rpc error: code = NotFound desc = could not find container \"19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508\": container with ID starting with 19548b383481c307bc20df6213e405d9cb2389dd2f8d2976be6e401c14505508 not found: ID does not exist" Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.577821 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-kf8zp"] Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.581781 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-kf8zp"] Dec 04 10:31:19 crc kubenswrapper[4943]: I1204 10:31:19.585078 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-q9s4x"] Dec 04 10:31:19 crc kubenswrapper[4943]: W1204 10:31:19.585755 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc076efe8_6293_4cd9_a27d_d6196e44b0e2.slice/crio-c837524f810ec82eed9b7b785475b2cf2f134fb1fcbc5721d20ad55a30479cc1 WatchSource:0}: Error finding container c837524f810ec82eed9b7b785475b2cf2f134fb1fcbc5721d20ad55a30479cc1: Status 404 returned error can't find the container with id c837524f810ec82eed9b7b785475b2cf2f134fb1fcbc5721d20ad55a30479cc1 Dec 04 10:31:20 crc kubenswrapper[4943]: I1204 10:31:20.526028 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-q9s4x" event={"ID":"c076efe8-6293-4cd9-a27d-d6196e44b0e2","Type":"ContainerStarted","Data":"c837524f810ec82eed9b7b785475b2cf2f134fb1fcbc5721d20ad55a30479cc1"} Dec 04 10:31:20 crc kubenswrapper[4943]: I1204 10:31:20.527883 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xvmdn" podUID="3602eccf-ee26-4836-a73b-15f8e65424b3" containerName="registry-server" containerID="cri-o://c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd" gracePeriod=2 Dec 04 10:31:20 crc kubenswrapper[4943]: I1204 10:31:20.570877 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46cb23ec-4521-4cff-aae6-3183a0c6b5bb" path="/var/lib/kubelet/pods/46cb23ec-4521-4cff-aae6-3183a0c6b5bb/volumes" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.390611 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.539949 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-q9s4x" event={"ID":"c076efe8-6293-4cd9-a27d-d6196e44b0e2","Type":"ContainerStarted","Data":"6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f"} Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.544092 4943 generic.go:334] "Generic (PLEG): container finished" podID="3602eccf-ee26-4836-a73b-15f8e65424b3" containerID="c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd" exitCode=0 Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.544139 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xvmdn" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.544160 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvmdn" event={"ID":"3602eccf-ee26-4836-a73b-15f8e65424b3","Type":"ContainerDied","Data":"c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd"} Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.544246 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xvmdn" event={"ID":"3602eccf-ee26-4836-a73b-15f8e65424b3","Type":"ContainerDied","Data":"4d8c0056a9370ce8f50ba20b487a47ad854460fc9d5484c3ca781a0f77fbc43d"} Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.544275 4943 scope.go:117] "RemoveContainer" containerID="c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.552476 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdvlj\" (UniqueName: \"kubernetes.io/projected/3602eccf-ee26-4836-a73b-15f8e65424b3-kube-api-access-zdvlj\") pod \"3602eccf-ee26-4836-a73b-15f8e65424b3\" (UID: \"3602eccf-ee26-4836-a73b-15f8e65424b3\") " Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.552581 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3602eccf-ee26-4836-a73b-15f8e65424b3-utilities\") pod \"3602eccf-ee26-4836-a73b-15f8e65424b3\" (UID: \"3602eccf-ee26-4836-a73b-15f8e65424b3\") " Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.552670 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3602eccf-ee26-4836-a73b-15f8e65424b3-catalog-content\") pod \"3602eccf-ee26-4836-a73b-15f8e65424b3\" (UID: \"3602eccf-ee26-4836-a73b-15f8e65424b3\") " Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.553816 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3602eccf-ee26-4836-a73b-15f8e65424b3-utilities" (OuterVolumeSpecName: "utilities") pod "3602eccf-ee26-4836-a73b-15f8e65424b3" (UID: "3602eccf-ee26-4836-a73b-15f8e65424b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.560108 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-q9s4x" podStartSLOduration=2.745680197 podStartE2EDuration="3.560070344s" podCreationTimestamp="2025-12-04 10:31:18 +0000 UTC" firstStartedPulling="2025-12-04 10:31:19.589950156 +0000 UTC m=+928.178926024" lastFinishedPulling="2025-12-04 10:31:20.404340303 +0000 UTC m=+928.993316171" observedRunningTime="2025-12-04 10:31:21.556301637 +0000 UTC m=+930.145277545" watchObservedRunningTime="2025-12-04 10:31:21.560070344 +0000 UTC m=+930.149046212" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.561758 4943 scope.go:117] "RemoveContainer" containerID="f54355e898831ce8d1b46f69563bb7d2dc0b129752fd423bd0c2622d3928fc7f" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.570866 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3602eccf-ee26-4836-a73b-15f8e65424b3-kube-api-access-zdvlj" (OuterVolumeSpecName: "kube-api-access-zdvlj") pod "3602eccf-ee26-4836-a73b-15f8e65424b3" (UID: "3602eccf-ee26-4836-a73b-15f8e65424b3"). InnerVolumeSpecName "kube-api-access-zdvlj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.588310 4943 scope.go:117] "RemoveContainer" containerID="342d4a2523389e7ea3aee0394afe1fc92d8c23e674a7cecbba50864d7c92b9f4" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.602557 4943 scope.go:117] "RemoveContainer" containerID="c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd" Dec 04 10:31:21 crc kubenswrapper[4943]: E1204 10:31:21.602989 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd\": container with ID starting with c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd not found: ID does not exist" containerID="c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.603042 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd"} err="failed to get container status \"c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd\": rpc error: code = NotFound desc = could not find container \"c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd\": container with ID starting with c58601cb5543bca7d851712c0d273d461f1dab216aff0fee7329368ea2d131bd not found: ID does not exist" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.603071 4943 scope.go:117] "RemoveContainer" containerID="f54355e898831ce8d1b46f69563bb7d2dc0b129752fd423bd0c2622d3928fc7f" Dec 04 10:31:21 crc kubenswrapper[4943]: E1204 10:31:21.603553 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f54355e898831ce8d1b46f69563bb7d2dc0b129752fd423bd0c2622d3928fc7f\": container with ID starting with f54355e898831ce8d1b46f69563bb7d2dc0b129752fd423bd0c2622d3928fc7f not found: ID does not exist" containerID="f54355e898831ce8d1b46f69563bb7d2dc0b129752fd423bd0c2622d3928fc7f" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.603605 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f54355e898831ce8d1b46f69563bb7d2dc0b129752fd423bd0c2622d3928fc7f"} err="failed to get container status \"f54355e898831ce8d1b46f69563bb7d2dc0b129752fd423bd0c2622d3928fc7f\": rpc error: code = NotFound desc = could not find container \"f54355e898831ce8d1b46f69563bb7d2dc0b129752fd423bd0c2622d3928fc7f\": container with ID starting with f54355e898831ce8d1b46f69563bb7d2dc0b129752fd423bd0c2622d3928fc7f not found: ID does not exist" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.603641 4943 scope.go:117] "RemoveContainer" containerID="342d4a2523389e7ea3aee0394afe1fc92d8c23e674a7cecbba50864d7c92b9f4" Dec 04 10:31:21 crc kubenswrapper[4943]: E1204 10:31:21.603946 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"342d4a2523389e7ea3aee0394afe1fc92d8c23e674a7cecbba50864d7c92b9f4\": container with ID starting with 342d4a2523389e7ea3aee0394afe1fc92d8c23e674a7cecbba50864d7c92b9f4 not found: ID does not exist" containerID="342d4a2523389e7ea3aee0394afe1fc92d8c23e674a7cecbba50864d7c92b9f4" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.603976 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"342d4a2523389e7ea3aee0394afe1fc92d8c23e674a7cecbba50864d7c92b9f4"} err="failed to get container status \"342d4a2523389e7ea3aee0394afe1fc92d8c23e674a7cecbba50864d7c92b9f4\": rpc error: code = NotFound desc = could not find container \"342d4a2523389e7ea3aee0394afe1fc92d8c23e674a7cecbba50864d7c92b9f4\": container with ID starting with 342d4a2523389e7ea3aee0394afe1fc92d8c23e674a7cecbba50864d7c92b9f4 not found: ID does not exist" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.609608 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3602eccf-ee26-4836-a73b-15f8e65424b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3602eccf-ee26-4836-a73b-15f8e65424b3" (UID: "3602eccf-ee26-4836-a73b-15f8e65424b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.653738 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdvlj\" (UniqueName: \"kubernetes.io/projected/3602eccf-ee26-4836-a73b-15f8e65424b3-kube-api-access-zdvlj\") on node \"crc\" DevicePath \"\"" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.653772 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3602eccf-ee26-4836-a73b-15f8e65424b3-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.653788 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3602eccf-ee26-4836-a73b-15f8e65424b3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.884923 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xvmdn"] Dec 04 10:31:21 crc kubenswrapper[4943]: I1204 10:31:21.889910 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xvmdn"] Dec 04 10:31:22 crc kubenswrapper[4943]: I1204 10:31:22.573787 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3602eccf-ee26-4836-a73b-15f8e65424b3" path="/var/lib/kubelet/pods/3602eccf-ee26-4836-a73b-15f8e65424b3/volumes" Dec 04 10:31:23 crc kubenswrapper[4943]: E1204 10:31:23.704792 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b156b3d_7929_43c5_9628_55a56b252794.slice/crio-06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b156b3d_7929_43c5_9628_55a56b252794.slice/crio-conmon-06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca.scope\": RecentStats: unable to find data in memory cache]" Dec 04 10:31:29 crc kubenswrapper[4943]: I1204 10:31:29.338105 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-q9s4x" Dec 04 10:31:29 crc kubenswrapper[4943]: I1204 10:31:29.338398 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-q9s4x" Dec 04 10:31:29 crc kubenswrapper[4943]: I1204 10:31:29.365989 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-q9s4x" Dec 04 10:31:29 crc kubenswrapper[4943]: I1204 10:31:29.612178 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-q9s4x" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.664559 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7"] Dec 04 10:31:32 crc kubenswrapper[4943]: E1204 10:31:32.664951 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3602eccf-ee26-4836-a73b-15f8e65424b3" containerName="extract-utilities" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.664998 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3602eccf-ee26-4836-a73b-15f8e65424b3" containerName="extract-utilities" Dec 04 10:31:32 crc kubenswrapper[4943]: E1204 10:31:32.665026 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46cb23ec-4521-4cff-aae6-3183a0c6b5bb" containerName="registry-server" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.665040 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="46cb23ec-4521-4cff-aae6-3183a0c6b5bb" containerName="registry-server" Dec 04 10:31:32 crc kubenswrapper[4943]: E1204 10:31:32.665068 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3602eccf-ee26-4836-a73b-15f8e65424b3" containerName="extract-content" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.665082 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3602eccf-ee26-4836-a73b-15f8e65424b3" containerName="extract-content" Dec 04 10:31:32 crc kubenswrapper[4943]: E1204 10:31:32.665105 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3602eccf-ee26-4836-a73b-15f8e65424b3" containerName="registry-server" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.665117 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3602eccf-ee26-4836-a73b-15f8e65424b3" containerName="registry-server" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.665403 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3602eccf-ee26-4836-a73b-15f8e65424b3" containerName="registry-server" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.665431 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="46cb23ec-4521-4cff-aae6-3183a0c6b5bb" containerName="registry-server" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.666484 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.669554 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6gzwg" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.676748 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7"] Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.702047 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ec088ed-72ed-4c49-bba9-53047c3bf094-bundle\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7\" (UID: \"5ec088ed-72ed-4c49-bba9-53047c3bf094\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.702156 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fk7s\" (UniqueName: \"kubernetes.io/projected/5ec088ed-72ed-4c49-bba9-53047c3bf094-kube-api-access-2fk7s\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7\" (UID: \"5ec088ed-72ed-4c49-bba9-53047c3bf094\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.702297 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ec088ed-72ed-4c49-bba9-53047c3bf094-util\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7\" (UID: \"5ec088ed-72ed-4c49-bba9-53047c3bf094\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.803928 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fk7s\" (UniqueName: \"kubernetes.io/projected/5ec088ed-72ed-4c49-bba9-53047c3bf094-kube-api-access-2fk7s\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7\" (UID: \"5ec088ed-72ed-4c49-bba9-53047c3bf094\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.804003 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ec088ed-72ed-4c49-bba9-53047c3bf094-util\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7\" (UID: \"5ec088ed-72ed-4c49-bba9-53047c3bf094\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.804150 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ec088ed-72ed-4c49-bba9-53047c3bf094-bundle\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7\" (UID: \"5ec088ed-72ed-4c49-bba9-53047c3bf094\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.805035 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ec088ed-72ed-4c49-bba9-53047c3bf094-bundle\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7\" (UID: \"5ec088ed-72ed-4c49-bba9-53047c3bf094\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.805336 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ec088ed-72ed-4c49-bba9-53047c3bf094-util\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7\" (UID: \"5ec088ed-72ed-4c49-bba9-53047c3bf094\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:32 crc kubenswrapper[4943]: I1204 10:31:32.839136 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fk7s\" (UniqueName: \"kubernetes.io/projected/5ec088ed-72ed-4c49-bba9-53047c3bf094-kube-api-access-2fk7s\") pod \"55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7\" (UID: \"5ec088ed-72ed-4c49-bba9-53047c3bf094\") " pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:33 crc kubenswrapper[4943]: I1204 10:31:33.066189 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:33 crc kubenswrapper[4943]: I1204 10:31:33.237656 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7"] Dec 04 10:31:33 crc kubenswrapper[4943]: I1204 10:31:33.617305 4943 generic.go:334] "Generic (PLEG): container finished" podID="5ec088ed-72ed-4c49-bba9-53047c3bf094" containerID="f0e6f2b6bd6291af317000838c52b98396866b3d0ad06a75314077430f652864" exitCode=0 Dec 04 10:31:33 crc kubenswrapper[4943]: I1204 10:31:33.617363 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" event={"ID":"5ec088ed-72ed-4c49-bba9-53047c3bf094","Type":"ContainerDied","Data":"f0e6f2b6bd6291af317000838c52b98396866b3d0ad06a75314077430f652864"} Dec 04 10:31:33 crc kubenswrapper[4943]: I1204 10:31:33.617394 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" event={"ID":"5ec088ed-72ed-4c49-bba9-53047c3bf094","Type":"ContainerStarted","Data":"26b84d2bafa8bfab48fe2fb9ff8945f507e71ba0f40891f95db7bba8e1005451"} Dec 04 10:31:33 crc kubenswrapper[4943]: E1204 10:31:33.853438 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b156b3d_7929_43c5_9628_55a56b252794.slice/crio-conmon-06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b156b3d_7929_43c5_9628_55a56b252794.slice/crio-06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca.scope\": RecentStats: unable to find data in memory cache]" Dec 04 10:31:34 crc kubenswrapper[4943]: I1204 10:31:34.627588 4943 generic.go:334] "Generic (PLEG): container finished" podID="5ec088ed-72ed-4c49-bba9-53047c3bf094" containerID="f8d98b25f4434d0a16b3ee37103940be8fff2ad1f838e3c4f5bc113dc67b3b14" exitCode=0 Dec 04 10:31:34 crc kubenswrapper[4943]: I1204 10:31:34.627642 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" event={"ID":"5ec088ed-72ed-4c49-bba9-53047c3bf094","Type":"ContainerDied","Data":"f8d98b25f4434d0a16b3ee37103940be8fff2ad1f838e3c4f5bc113dc67b3b14"} Dec 04 10:31:35 crc kubenswrapper[4943]: I1204 10:31:35.641272 4943 generic.go:334] "Generic (PLEG): container finished" podID="5ec088ed-72ed-4c49-bba9-53047c3bf094" containerID="18c2e1103834644ee7013475e9f5e7580ba6c9948857a4ef3c560c850bf7c70f" exitCode=0 Dec 04 10:31:35 crc kubenswrapper[4943]: I1204 10:31:35.641388 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" event={"ID":"5ec088ed-72ed-4c49-bba9-53047c3bf094","Type":"ContainerDied","Data":"18c2e1103834644ee7013475e9f5e7580ba6c9948857a4ef3c560c850bf7c70f"} Dec 04 10:31:36 crc kubenswrapper[4943]: I1204 10:31:36.929479 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:37 crc kubenswrapper[4943]: I1204 10:31:37.045676 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ec088ed-72ed-4c49-bba9-53047c3bf094-bundle\") pod \"5ec088ed-72ed-4c49-bba9-53047c3bf094\" (UID: \"5ec088ed-72ed-4c49-bba9-53047c3bf094\") " Dec 04 10:31:37 crc kubenswrapper[4943]: I1204 10:31:37.045828 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fk7s\" (UniqueName: \"kubernetes.io/projected/5ec088ed-72ed-4c49-bba9-53047c3bf094-kube-api-access-2fk7s\") pod \"5ec088ed-72ed-4c49-bba9-53047c3bf094\" (UID: \"5ec088ed-72ed-4c49-bba9-53047c3bf094\") " Dec 04 10:31:37 crc kubenswrapper[4943]: I1204 10:31:37.045863 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ec088ed-72ed-4c49-bba9-53047c3bf094-util\") pod \"5ec088ed-72ed-4c49-bba9-53047c3bf094\" (UID: \"5ec088ed-72ed-4c49-bba9-53047c3bf094\") " Dec 04 10:31:37 crc kubenswrapper[4943]: I1204 10:31:37.046774 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ec088ed-72ed-4c49-bba9-53047c3bf094-bundle" (OuterVolumeSpecName: "bundle") pod "5ec088ed-72ed-4c49-bba9-53047c3bf094" (UID: "5ec088ed-72ed-4c49-bba9-53047c3bf094"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:31:37 crc kubenswrapper[4943]: I1204 10:31:37.055173 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ec088ed-72ed-4c49-bba9-53047c3bf094-kube-api-access-2fk7s" (OuterVolumeSpecName: "kube-api-access-2fk7s") pod "5ec088ed-72ed-4c49-bba9-53047c3bf094" (UID: "5ec088ed-72ed-4c49-bba9-53047c3bf094"). InnerVolumeSpecName "kube-api-access-2fk7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:31:37 crc kubenswrapper[4943]: I1204 10:31:37.064043 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ec088ed-72ed-4c49-bba9-53047c3bf094-util" (OuterVolumeSpecName: "util") pod "5ec088ed-72ed-4c49-bba9-53047c3bf094" (UID: "5ec088ed-72ed-4c49-bba9-53047c3bf094"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:31:37 crc kubenswrapper[4943]: I1204 10:31:37.147497 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fk7s\" (UniqueName: \"kubernetes.io/projected/5ec088ed-72ed-4c49-bba9-53047c3bf094-kube-api-access-2fk7s\") on node \"crc\" DevicePath \"\"" Dec 04 10:31:37 crc kubenswrapper[4943]: I1204 10:31:37.147537 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ec088ed-72ed-4c49-bba9-53047c3bf094-util\") on node \"crc\" DevicePath \"\"" Dec 04 10:31:37 crc kubenswrapper[4943]: I1204 10:31:37.147554 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ec088ed-72ed-4c49-bba9-53047c3bf094-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:31:37 crc kubenswrapper[4943]: I1204 10:31:37.661609 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" event={"ID":"5ec088ed-72ed-4c49-bba9-53047c3bf094","Type":"ContainerDied","Data":"26b84d2bafa8bfab48fe2fb9ff8945f507e71ba0f40891f95db7bba8e1005451"} Dec 04 10:31:37 crc kubenswrapper[4943]: I1204 10:31:37.661716 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26b84d2bafa8bfab48fe2fb9ff8945f507e71ba0f40891f95db7bba8e1005451" Dec 04 10:31:37 crc kubenswrapper[4943]: I1204 10:31:37.661832 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7" Dec 04 10:31:44 crc kubenswrapper[4943]: E1204 10:31:44.005305 4943 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b156b3d_7929_43c5_9628_55a56b252794.slice/crio-06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b156b3d_7929_43c5_9628_55a56b252794.slice/crio-conmon-06d5df8cb7f31bda808c48fe5fd9f7aa91ace4ccb4f0ee29fb5cba364879a8ca.scope\": RecentStats: unable to find data in memory cache]" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.686038 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7"] Dec 04 10:31:45 crc kubenswrapper[4943]: E1204 10:31:45.686530 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec088ed-72ed-4c49-bba9-53047c3bf094" containerName="util" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.686546 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec088ed-72ed-4c49-bba9-53047c3bf094" containerName="util" Dec 04 10:31:45 crc kubenswrapper[4943]: E1204 10:31:45.686567 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec088ed-72ed-4c49-bba9-53047c3bf094" containerName="extract" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.686575 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec088ed-72ed-4c49-bba9-53047c3bf094" containerName="extract" Dec 04 10:31:45 crc kubenswrapper[4943]: E1204 10:31:45.686600 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec088ed-72ed-4c49-bba9-53047c3bf094" containerName="pull" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.686610 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec088ed-72ed-4c49-bba9-53047c3bf094" containerName="pull" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.686885 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ec088ed-72ed-4c49-bba9-53047c3bf094" containerName="extract" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.687603 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.692899 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-np29h" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.693211 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.693696 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.703485 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7"] Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.875559 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-apiservice-cert\") pod \"mariadb-operator-controller-manager-84c9b4cc96-2frr7\" (UID: \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\") " pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.875602 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5t6r\" (UniqueName: \"kubernetes.io/projected/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-kube-api-access-l5t6r\") pod \"mariadb-operator-controller-manager-84c9b4cc96-2frr7\" (UID: \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\") " pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.875636 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-webhook-cert\") pod \"mariadb-operator-controller-manager-84c9b4cc96-2frr7\" (UID: \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\") " pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.977107 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-apiservice-cert\") pod \"mariadb-operator-controller-manager-84c9b4cc96-2frr7\" (UID: \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\") " pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.977193 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5t6r\" (UniqueName: \"kubernetes.io/projected/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-kube-api-access-l5t6r\") pod \"mariadb-operator-controller-manager-84c9b4cc96-2frr7\" (UID: \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\") " pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.977305 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-webhook-cert\") pod \"mariadb-operator-controller-manager-84c9b4cc96-2frr7\" (UID: \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\") " pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.983311 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-apiservice-cert\") pod \"mariadb-operator-controller-manager-84c9b4cc96-2frr7\" (UID: \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\") " pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.988487 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-webhook-cert\") pod \"mariadb-operator-controller-manager-84c9b4cc96-2frr7\" (UID: \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\") " pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:45 crc kubenswrapper[4943]: I1204 10:31:45.994902 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5t6r\" (UniqueName: \"kubernetes.io/projected/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-kube-api-access-l5t6r\") pod \"mariadb-operator-controller-manager-84c9b4cc96-2frr7\" (UID: \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\") " pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:46 crc kubenswrapper[4943]: I1204 10:31:46.023448 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:46 crc kubenswrapper[4943]: I1204 10:31:46.245267 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7"] Dec 04 10:31:46 crc kubenswrapper[4943]: I1204 10:31:46.731856 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" event={"ID":"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0","Type":"ContainerStarted","Data":"21322eb6f4d98f29386068d4fdc46fb089cc0ff5dc166c109affb7a6c3e45828"} Dec 04 10:31:48 crc kubenswrapper[4943]: I1204 10:31:48.629377 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:31:48 crc kubenswrapper[4943]: I1204 10:31:48.629737 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:31:49 crc kubenswrapper[4943]: I1204 10:31:49.754859 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" event={"ID":"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0","Type":"ContainerStarted","Data":"afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f"} Dec 04 10:31:49 crc kubenswrapper[4943]: I1204 10:31:49.755209 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:49 crc kubenswrapper[4943]: I1204 10:31:49.776910 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" podStartSLOduration=1.5328587599999999 podStartE2EDuration="4.776882502s" podCreationTimestamp="2025-12-04 10:31:45 +0000 UTC" firstStartedPulling="2025-12-04 10:31:46.260862292 +0000 UTC m=+954.849838160" lastFinishedPulling="2025-12-04 10:31:49.504886034 +0000 UTC m=+958.093861902" observedRunningTime="2025-12-04 10:31:49.77260074 +0000 UTC m=+958.361576628" watchObservedRunningTime="2025-12-04 10:31:49.776882502 +0000 UTC m=+958.365858410" Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.141311 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xxzth"] Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.142978 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.148874 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxzth"] Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.334679 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3644228f-29de-46a5-be32-cea19ae03106-utilities\") pod \"redhat-marketplace-xxzth\" (UID: \"3644228f-29de-46a5-be32-cea19ae03106\") " pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.334764 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3644228f-29de-46a5-be32-cea19ae03106-catalog-content\") pod \"redhat-marketplace-xxzth\" (UID: \"3644228f-29de-46a5-be32-cea19ae03106\") " pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.335014 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b69m2\" (UniqueName: \"kubernetes.io/projected/3644228f-29de-46a5-be32-cea19ae03106-kube-api-access-b69m2\") pod \"redhat-marketplace-xxzth\" (UID: \"3644228f-29de-46a5-be32-cea19ae03106\") " pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.436684 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3644228f-29de-46a5-be32-cea19ae03106-utilities\") pod \"redhat-marketplace-xxzth\" (UID: \"3644228f-29de-46a5-be32-cea19ae03106\") " pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.436752 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3644228f-29de-46a5-be32-cea19ae03106-catalog-content\") pod \"redhat-marketplace-xxzth\" (UID: \"3644228f-29de-46a5-be32-cea19ae03106\") " pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.436804 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b69m2\" (UniqueName: \"kubernetes.io/projected/3644228f-29de-46a5-be32-cea19ae03106-kube-api-access-b69m2\") pod \"redhat-marketplace-xxzth\" (UID: \"3644228f-29de-46a5-be32-cea19ae03106\") " pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.437277 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3644228f-29de-46a5-be32-cea19ae03106-utilities\") pod \"redhat-marketplace-xxzth\" (UID: \"3644228f-29de-46a5-be32-cea19ae03106\") " pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.437452 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3644228f-29de-46a5-be32-cea19ae03106-catalog-content\") pod \"redhat-marketplace-xxzth\" (UID: \"3644228f-29de-46a5-be32-cea19ae03106\") " pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.455669 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b69m2\" (UniqueName: \"kubernetes.io/projected/3644228f-29de-46a5-be32-cea19ae03106-kube-api-access-b69m2\") pod \"redhat-marketplace-xxzth\" (UID: \"3644228f-29de-46a5-be32-cea19ae03106\") " pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.461658 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.690784 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxzth"] Dec 04 10:31:50 crc kubenswrapper[4943]: W1204 10:31:50.705435 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3644228f_29de_46a5_be32_cea19ae03106.slice/crio-099af4fab2505f3533981c86cc76352cfa9348d8021479733397b93898d72046 WatchSource:0}: Error finding container 099af4fab2505f3533981c86cc76352cfa9348d8021479733397b93898d72046: Status 404 returned error can't find the container with id 099af4fab2505f3533981c86cc76352cfa9348d8021479733397b93898d72046 Dec 04 10:31:50 crc kubenswrapper[4943]: I1204 10:31:50.764925 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxzth" event={"ID":"3644228f-29de-46a5-be32-cea19ae03106","Type":"ContainerStarted","Data":"099af4fab2505f3533981c86cc76352cfa9348d8021479733397b93898d72046"} Dec 04 10:31:51 crc kubenswrapper[4943]: I1204 10:31:51.771461 4943 generic.go:334] "Generic (PLEG): container finished" podID="3644228f-29de-46a5-be32-cea19ae03106" containerID="8192b758f374d36747813711d002d2b42afd3342f1cfcee151a4900d6d31aade" exitCode=0 Dec 04 10:31:51 crc kubenswrapper[4943]: I1204 10:31:51.772068 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxzth" event={"ID":"3644228f-29de-46a5-be32-cea19ae03106","Type":"ContainerDied","Data":"8192b758f374d36747813711d002d2b42afd3342f1cfcee151a4900d6d31aade"} Dec 04 10:31:52 crc kubenswrapper[4943]: E1204 10:31:52.469660 4943 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/14979ec867047041bf2b6da846ef479ef0506da222ae5885f6454c1bd20042b5/diff" to get inode usage: stat /var/lib/containers/storage/overlay/14979ec867047041bf2b6da846ef479ef0506da222ae5885f6454c1bd20042b5/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openshift-marketplace_certified-operators-t2qqp_5b156b3d-7929-43c5-9628-55a56b252794/extract-content/0.log" to get inode usage: stat /var/log/pods/openshift-marketplace_certified-operators-t2qqp_5b156b3d-7929-43c5-9628-55a56b252794/extract-content/0.log: no such file or directory Dec 04 10:31:53 crc kubenswrapper[4943]: I1204 10:31:53.786347 4943 generic.go:334] "Generic (PLEG): container finished" podID="3644228f-29de-46a5-be32-cea19ae03106" containerID="17f4271b79c8061fba803f25e0527385063662296ba4a41fd6b4f9897b20b491" exitCode=0 Dec 04 10:31:53 crc kubenswrapper[4943]: I1204 10:31:53.786679 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxzth" event={"ID":"3644228f-29de-46a5-be32-cea19ae03106","Type":"ContainerDied","Data":"17f4271b79c8061fba803f25e0527385063662296ba4a41fd6b4f9897b20b491"} Dec 04 10:31:54 crc kubenswrapper[4943]: I1204 10:31:54.793587 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxzth" event={"ID":"3644228f-29de-46a5-be32-cea19ae03106","Type":"ContainerStarted","Data":"3ab38e95ba26097b632d6a769a36b001badc2365f77cb6d7175dd69d9f026458"} Dec 04 10:31:54 crc kubenswrapper[4943]: I1204 10:31:54.814188 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xxzth" podStartSLOduration=2.363285037 podStartE2EDuration="4.814164985s" podCreationTimestamp="2025-12-04 10:31:50 +0000 UTC" firstStartedPulling="2025-12-04 10:31:51.773221004 +0000 UTC m=+960.362196872" lastFinishedPulling="2025-12-04 10:31:54.224100952 +0000 UTC m=+962.813076820" observedRunningTime="2025-12-04 10:31:54.810848571 +0000 UTC m=+963.399824469" watchObservedRunningTime="2025-12-04 10:31:54.814164985 +0000 UTC m=+963.403140853" Dec 04 10:31:56 crc kubenswrapper[4943]: I1204 10:31:56.027495 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:31:59 crc kubenswrapper[4943]: I1204 10:31:59.528148 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-9d45w"] Dec 04 10:31:59 crc kubenswrapper[4943]: I1204 10:31:59.529077 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-9d45w" Dec 04 10:31:59 crc kubenswrapper[4943]: I1204 10:31:59.531277 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-pqhrg" Dec 04 10:31:59 crc kubenswrapper[4943]: I1204 10:31:59.542080 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-9d45w"] Dec 04 10:31:59 crc kubenswrapper[4943]: I1204 10:31:59.553502 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p52m\" (UniqueName: \"kubernetes.io/projected/9ee82f5b-e5aa-46cd-929a-6e9f32de2a43-kube-api-access-4p52m\") pod \"infra-operator-index-9d45w\" (UID: \"9ee82f5b-e5aa-46cd-929a-6e9f32de2a43\") " pod="openstack-operators/infra-operator-index-9d45w" Dec 04 10:31:59 crc kubenswrapper[4943]: I1204 10:31:59.654576 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p52m\" (UniqueName: \"kubernetes.io/projected/9ee82f5b-e5aa-46cd-929a-6e9f32de2a43-kube-api-access-4p52m\") pod \"infra-operator-index-9d45w\" (UID: \"9ee82f5b-e5aa-46cd-929a-6e9f32de2a43\") " pod="openstack-operators/infra-operator-index-9d45w" Dec 04 10:31:59 crc kubenswrapper[4943]: I1204 10:31:59.684004 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p52m\" (UniqueName: \"kubernetes.io/projected/9ee82f5b-e5aa-46cd-929a-6e9f32de2a43-kube-api-access-4p52m\") pod \"infra-operator-index-9d45w\" (UID: \"9ee82f5b-e5aa-46cd-929a-6e9f32de2a43\") " pod="openstack-operators/infra-operator-index-9d45w" Dec 04 10:31:59 crc kubenswrapper[4943]: I1204 10:31:59.844588 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-9d45w" Dec 04 10:32:00 crc kubenswrapper[4943]: I1204 10:32:00.247233 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-9d45w"] Dec 04 10:32:00 crc kubenswrapper[4943]: I1204 10:32:00.461892 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:32:00 crc kubenswrapper[4943]: I1204 10:32:00.461937 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:32:00 crc kubenswrapper[4943]: I1204 10:32:00.503449 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:32:00 crc kubenswrapper[4943]: I1204 10:32:00.826376 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-9d45w" event={"ID":"9ee82f5b-e5aa-46cd-929a-6e9f32de2a43","Type":"ContainerStarted","Data":"da16e087f0046a0b833ca9e3e1c557ccc80b2c54fa3a75554797e39bd4253698"} Dec 04 10:32:00 crc kubenswrapper[4943]: I1204 10:32:00.865407 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:32:04 crc kubenswrapper[4943]: I1204 10:32:04.927016 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-9d45w"] Dec 04 10:32:05 crc kubenswrapper[4943]: I1204 10:32:05.326078 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxzth"] Dec 04 10:32:05 crc kubenswrapper[4943]: I1204 10:32:05.326620 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xxzth" podUID="3644228f-29de-46a5-be32-cea19ae03106" containerName="registry-server" containerID="cri-o://3ab38e95ba26097b632d6a769a36b001badc2365f77cb6d7175dd69d9f026458" gracePeriod=2 Dec 04 10:32:05 crc kubenswrapper[4943]: I1204 10:32:05.730395 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-szxbw"] Dec 04 10:32:05 crc kubenswrapper[4943]: I1204 10:32:05.731074 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-szxbw" Dec 04 10:32:05 crc kubenswrapper[4943]: I1204 10:32:05.747244 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-szxbw"] Dec 04 10:32:05 crc kubenswrapper[4943]: I1204 10:32:05.846231 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5xqp\" (UniqueName: \"kubernetes.io/projected/56e01120-0537-47f3-92d0-c9b6ebc9651a-kube-api-access-w5xqp\") pod \"infra-operator-index-szxbw\" (UID: \"56e01120-0537-47f3-92d0-c9b6ebc9651a\") " pod="openstack-operators/infra-operator-index-szxbw" Dec 04 10:32:05 crc kubenswrapper[4943]: I1204 10:32:05.866814 4943 generic.go:334] "Generic (PLEG): container finished" podID="3644228f-29de-46a5-be32-cea19ae03106" containerID="3ab38e95ba26097b632d6a769a36b001badc2365f77cb6d7175dd69d9f026458" exitCode=0 Dec 04 10:32:05 crc kubenswrapper[4943]: I1204 10:32:05.866856 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxzth" event={"ID":"3644228f-29de-46a5-be32-cea19ae03106","Type":"ContainerDied","Data":"3ab38e95ba26097b632d6a769a36b001badc2365f77cb6d7175dd69d9f026458"} Dec 04 10:32:05 crc kubenswrapper[4943]: I1204 10:32:05.948489 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5xqp\" (UniqueName: \"kubernetes.io/projected/56e01120-0537-47f3-92d0-c9b6ebc9651a-kube-api-access-w5xqp\") pod \"infra-operator-index-szxbw\" (UID: \"56e01120-0537-47f3-92d0-c9b6ebc9651a\") " pod="openstack-operators/infra-operator-index-szxbw" Dec 04 10:32:05 crc kubenswrapper[4943]: I1204 10:32:05.967623 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5xqp\" (UniqueName: \"kubernetes.io/projected/56e01120-0537-47f3-92d0-c9b6ebc9651a-kube-api-access-w5xqp\") pod \"infra-operator-index-szxbw\" (UID: \"56e01120-0537-47f3-92d0-c9b6ebc9651a\") " pod="openstack-operators/infra-operator-index-szxbw" Dec 04 10:32:06 crc kubenswrapper[4943]: I1204 10:32:06.051458 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-szxbw" Dec 04 10:32:08 crc kubenswrapper[4943]: I1204 10:32:08.837960 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:32:08 crc kubenswrapper[4943]: I1204 10:32:08.879043 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-szxbw"] Dec 04 10:32:08 crc kubenswrapper[4943]: I1204 10:32:08.888513 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xxzth" event={"ID":"3644228f-29de-46a5-be32-cea19ae03106","Type":"ContainerDied","Data":"099af4fab2505f3533981c86cc76352cfa9348d8021479733397b93898d72046"} Dec 04 10:32:08 crc kubenswrapper[4943]: I1204 10:32:08.888565 4943 scope.go:117] "RemoveContainer" containerID="3ab38e95ba26097b632d6a769a36b001badc2365f77cb6d7175dd69d9f026458" Dec 04 10:32:08 crc kubenswrapper[4943]: I1204 10:32:08.888643 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xxzth" Dec 04 10:32:08 crc kubenswrapper[4943]: W1204 10:32:08.893541 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56e01120_0537_47f3_92d0_c9b6ebc9651a.slice/crio-771be2b6914a587f236deb123a86642711698a7330a333bee1a982e41d1e2e67 WatchSource:0}: Error finding container 771be2b6914a587f236deb123a86642711698a7330a333bee1a982e41d1e2e67: Status 404 returned error can't find the container with id 771be2b6914a587f236deb123a86642711698a7330a333bee1a982e41d1e2e67 Dec 04 10:32:08 crc kubenswrapper[4943]: I1204 10:32:08.905895 4943 scope.go:117] "RemoveContainer" containerID="17f4271b79c8061fba803f25e0527385063662296ba4a41fd6b4f9897b20b491" Dec 04 10:32:08 crc kubenswrapper[4943]: I1204 10:32:08.928395 4943 scope.go:117] "RemoveContainer" containerID="8192b758f374d36747813711d002d2b42afd3342f1cfcee151a4900d6d31aade" Dec 04 10:32:08 crc kubenswrapper[4943]: I1204 10:32:08.999435 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b69m2\" (UniqueName: \"kubernetes.io/projected/3644228f-29de-46a5-be32-cea19ae03106-kube-api-access-b69m2\") pod \"3644228f-29de-46a5-be32-cea19ae03106\" (UID: \"3644228f-29de-46a5-be32-cea19ae03106\") " Dec 04 10:32:08 crc kubenswrapper[4943]: I1204 10:32:08.999527 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3644228f-29de-46a5-be32-cea19ae03106-catalog-content\") pod \"3644228f-29de-46a5-be32-cea19ae03106\" (UID: \"3644228f-29de-46a5-be32-cea19ae03106\") " Dec 04 10:32:08 crc kubenswrapper[4943]: I1204 10:32:08.999635 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3644228f-29de-46a5-be32-cea19ae03106-utilities\") pod \"3644228f-29de-46a5-be32-cea19ae03106\" (UID: \"3644228f-29de-46a5-be32-cea19ae03106\") " Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.000692 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3644228f-29de-46a5-be32-cea19ae03106-utilities" (OuterVolumeSpecName: "utilities") pod "3644228f-29de-46a5-be32-cea19ae03106" (UID: "3644228f-29de-46a5-be32-cea19ae03106"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.004952 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3644228f-29de-46a5-be32-cea19ae03106-kube-api-access-b69m2" (OuterVolumeSpecName: "kube-api-access-b69m2") pod "3644228f-29de-46a5-be32-cea19ae03106" (UID: "3644228f-29de-46a5-be32-cea19ae03106"). InnerVolumeSpecName "kube-api-access-b69m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.018581 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3644228f-29de-46a5-be32-cea19ae03106-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3644228f-29de-46a5-be32-cea19ae03106" (UID: "3644228f-29de-46a5-be32-cea19ae03106"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.101168 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3644228f-29de-46a5-be32-cea19ae03106-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.101229 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b69m2\" (UniqueName: \"kubernetes.io/projected/3644228f-29de-46a5-be32-cea19ae03106-kube-api-access-b69m2\") on node \"crc\" DevicePath \"\"" Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.101241 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3644228f-29de-46a5-be32-cea19ae03106-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.226973 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxzth"] Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.231481 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xxzth"] Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.895402 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-9d45w" event={"ID":"9ee82f5b-e5aa-46cd-929a-6e9f32de2a43","Type":"ContainerStarted","Data":"665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac"} Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.895478 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-9d45w" podUID="9ee82f5b-e5aa-46cd-929a-6e9f32de2a43" containerName="registry-server" containerID="cri-o://665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac" gracePeriod=2 Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.897533 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-szxbw" event={"ID":"56e01120-0537-47f3-92d0-c9b6ebc9651a","Type":"ContainerStarted","Data":"5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f"} Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.897599 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-szxbw" event={"ID":"56e01120-0537-47f3-92d0-c9b6ebc9651a","Type":"ContainerStarted","Data":"771be2b6914a587f236deb123a86642711698a7330a333bee1a982e41d1e2e67"} Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.914153 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-9d45w" podStartSLOduration=2.373127375 podStartE2EDuration="10.914134108s" podCreationTimestamp="2025-12-04 10:31:59 +0000 UTC" firstStartedPulling="2025-12-04 10:32:00.257276462 +0000 UTC m=+968.846252330" lastFinishedPulling="2025-12-04 10:32:08.798283195 +0000 UTC m=+977.387259063" observedRunningTime="2025-12-04 10:32:09.912629306 +0000 UTC m=+978.501605174" watchObservedRunningTime="2025-12-04 10:32:09.914134108 +0000 UTC m=+978.503109976" Dec 04 10:32:09 crc kubenswrapper[4943]: I1204 10:32:09.933837 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-szxbw" podStartSLOduration=4.384382683 podStartE2EDuration="4.933811491s" podCreationTimestamp="2025-12-04 10:32:05 +0000 UTC" firstStartedPulling="2025-12-04 10:32:08.897534593 +0000 UTC m=+977.486510461" lastFinishedPulling="2025-12-04 10:32:09.446963401 +0000 UTC m=+978.035939269" observedRunningTime="2025-12-04 10:32:09.932034461 +0000 UTC m=+978.521010349" watchObservedRunningTime="2025-12-04 10:32:09.933811491 +0000 UTC m=+978.522787359" Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.245433 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-9d45w" Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.418473 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p52m\" (UniqueName: \"kubernetes.io/projected/9ee82f5b-e5aa-46cd-929a-6e9f32de2a43-kube-api-access-4p52m\") pod \"9ee82f5b-e5aa-46cd-929a-6e9f32de2a43\" (UID: \"9ee82f5b-e5aa-46cd-929a-6e9f32de2a43\") " Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.423987 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ee82f5b-e5aa-46cd-929a-6e9f32de2a43-kube-api-access-4p52m" (OuterVolumeSpecName: "kube-api-access-4p52m") pod "9ee82f5b-e5aa-46cd-929a-6e9f32de2a43" (UID: "9ee82f5b-e5aa-46cd-929a-6e9f32de2a43"). InnerVolumeSpecName "kube-api-access-4p52m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.519956 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p52m\" (UniqueName: \"kubernetes.io/projected/9ee82f5b-e5aa-46cd-929a-6e9f32de2a43-kube-api-access-4p52m\") on node \"crc\" DevicePath \"\"" Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.571587 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3644228f-29de-46a5-be32-cea19ae03106" path="/var/lib/kubelet/pods/3644228f-29de-46a5-be32-cea19ae03106/volumes" Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.907699 4943 generic.go:334] "Generic (PLEG): container finished" podID="9ee82f5b-e5aa-46cd-929a-6e9f32de2a43" containerID="665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac" exitCode=0 Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.907754 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-9d45w" event={"ID":"9ee82f5b-e5aa-46cd-929a-6e9f32de2a43","Type":"ContainerDied","Data":"665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac"} Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.907794 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-9d45w" Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.907837 4943 scope.go:117] "RemoveContainer" containerID="665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac" Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.907816 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-9d45w" event={"ID":"9ee82f5b-e5aa-46cd-929a-6e9f32de2a43","Type":"ContainerDied","Data":"da16e087f0046a0b833ca9e3e1c557ccc80b2c54fa3a75554797e39bd4253698"} Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.930716 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-9d45w"] Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.931290 4943 scope.go:117] "RemoveContainer" containerID="665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac" Dec 04 10:32:10 crc kubenswrapper[4943]: E1204 10:32:10.931710 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac\": container with ID starting with 665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac not found: ID does not exist" containerID="665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac" Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.931745 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac"} err="failed to get container status \"665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac\": rpc error: code = NotFound desc = could not find container \"665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac\": container with ID starting with 665ced8ae79fbf1e9626b4adef589dc0db0de256198ab72544f9a896911f1aac not found: ID does not exist" Dec 04 10:32:10 crc kubenswrapper[4943]: I1204 10:32:10.934635 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-9d45w"] Dec 04 10:32:12 crc kubenswrapper[4943]: I1204 10:32:12.573669 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ee82f5b-e5aa-46cd-929a-6e9f32de2a43" path="/var/lib/kubelet/pods/9ee82f5b-e5aa-46cd-929a-6e9f32de2a43/volumes" Dec 04 10:32:16 crc kubenswrapper[4943]: I1204 10:32:16.051829 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-szxbw" Dec 04 10:32:16 crc kubenswrapper[4943]: I1204 10:32:16.051900 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-szxbw" Dec 04 10:32:16 crc kubenswrapper[4943]: I1204 10:32:16.078719 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-szxbw" Dec 04 10:32:16 crc kubenswrapper[4943]: I1204 10:32:16.966085 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-szxbw" Dec 04 10:32:18 crc kubenswrapper[4943]: I1204 10:32:18.629713 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:32:18 crc kubenswrapper[4943]: I1204 10:32:18.629765 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.389166 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd"] Dec 04 10:32:31 crc kubenswrapper[4943]: E1204 10:32:31.390056 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ee82f5b-e5aa-46cd-929a-6e9f32de2a43" containerName="registry-server" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.390075 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ee82f5b-e5aa-46cd-929a-6e9f32de2a43" containerName="registry-server" Dec 04 10:32:31 crc kubenswrapper[4943]: E1204 10:32:31.390091 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3644228f-29de-46a5-be32-cea19ae03106" containerName="extract-utilities" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.390099 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3644228f-29de-46a5-be32-cea19ae03106" containerName="extract-utilities" Dec 04 10:32:31 crc kubenswrapper[4943]: E1204 10:32:31.390108 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3644228f-29de-46a5-be32-cea19ae03106" containerName="extract-content" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.390117 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3644228f-29de-46a5-be32-cea19ae03106" containerName="extract-content" Dec 04 10:32:31 crc kubenswrapper[4943]: E1204 10:32:31.390126 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3644228f-29de-46a5-be32-cea19ae03106" containerName="registry-server" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.390134 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3644228f-29de-46a5-be32-cea19ae03106" containerName="registry-server" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.390293 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3644228f-29de-46a5-be32-cea19ae03106" containerName="registry-server" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.390307 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ee82f5b-e5aa-46cd-929a-6e9f32de2a43" containerName="registry-server" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.391413 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.393220 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6gzwg" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.402239 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd"] Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.506131 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8cdr\" (UniqueName: \"kubernetes.io/projected/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-kube-api-access-j8cdr\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd\" (UID: \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.506229 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd\" (UID: \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.506265 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd\" (UID: \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.607360 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd\" (UID: \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.607409 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd\" (UID: \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.607491 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8cdr\" (UniqueName: \"kubernetes.io/projected/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-kube-api-access-j8cdr\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd\" (UID: \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.607930 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd\" (UID: \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.608027 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd\" (UID: \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.624691 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8cdr\" (UniqueName: \"kubernetes.io/projected/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-kube-api-access-j8cdr\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd\" (UID: \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:31 crc kubenswrapper[4943]: I1204 10:32:31.708189 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:32 crc kubenswrapper[4943]: I1204 10:32:32.118911 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd"] Dec 04 10:32:33 crc kubenswrapper[4943]: I1204 10:32:33.039936 4943 generic.go:334] "Generic (PLEG): container finished" podID="21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" containerID="58725c2bb3322200f2389af365db07f9329eda286b3c4c33901f7fc2b03df294" exitCode=0 Dec 04 10:32:33 crc kubenswrapper[4943]: I1204 10:32:33.039992 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" event={"ID":"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe","Type":"ContainerDied","Data":"58725c2bb3322200f2389af365db07f9329eda286b3c4c33901f7fc2b03df294"} Dec 04 10:32:33 crc kubenswrapper[4943]: I1204 10:32:33.040355 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" event={"ID":"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe","Type":"ContainerStarted","Data":"610c50dacec3f35e6c65291a668605f4c6b3f3877a33cadb3bb4dd4e24246b59"} Dec 04 10:32:35 crc kubenswrapper[4943]: I1204 10:32:35.054510 4943 generic.go:334] "Generic (PLEG): container finished" podID="21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" containerID="3e062fbd9e2ccd9b2c36ca08fd0945e54a464bd590527267dbccd2d508c6f1ea" exitCode=0 Dec 04 10:32:35 crc kubenswrapper[4943]: I1204 10:32:35.054821 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" event={"ID":"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe","Type":"ContainerDied","Data":"3e062fbd9e2ccd9b2c36ca08fd0945e54a464bd590527267dbccd2d508c6f1ea"} Dec 04 10:32:36 crc kubenswrapper[4943]: I1204 10:32:36.062186 4943 generic.go:334] "Generic (PLEG): container finished" podID="21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" containerID="6843515e1dead6038699866c979534ba43c1d0a8a3332005df7c21090dbdd13c" exitCode=0 Dec 04 10:32:36 crc kubenswrapper[4943]: I1204 10:32:36.062264 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" event={"ID":"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe","Type":"ContainerDied","Data":"6843515e1dead6038699866c979534ba43c1d0a8a3332005df7c21090dbdd13c"} Dec 04 10:32:37 crc kubenswrapper[4943]: I1204 10:32:37.515737 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:37 crc kubenswrapper[4943]: I1204 10:32:37.586659 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-bundle\") pod \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\" (UID: \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\") " Dec 04 10:32:37 crc kubenswrapper[4943]: I1204 10:32:37.586779 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-util\") pod \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\" (UID: \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\") " Dec 04 10:32:37 crc kubenswrapper[4943]: I1204 10:32:37.586851 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8cdr\" (UniqueName: \"kubernetes.io/projected/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-kube-api-access-j8cdr\") pod \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\" (UID: \"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe\") " Dec 04 10:32:37 crc kubenswrapper[4943]: I1204 10:32:37.588193 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-bundle" (OuterVolumeSpecName: "bundle") pod "21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" (UID: "21b39ac1-11f1-4bf2-bdd8-6e93062ecafe"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:32:37 crc kubenswrapper[4943]: I1204 10:32:37.593631 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-kube-api-access-j8cdr" (OuterVolumeSpecName: "kube-api-access-j8cdr") pod "21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" (UID: "21b39ac1-11f1-4bf2-bdd8-6e93062ecafe"). InnerVolumeSpecName "kube-api-access-j8cdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:32:37 crc kubenswrapper[4943]: I1204 10:32:37.601596 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-util" (OuterVolumeSpecName: "util") pod "21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" (UID: "21b39ac1-11f1-4bf2-bdd8-6e93062ecafe"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:32:37 crc kubenswrapper[4943]: I1204 10:32:37.688406 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-util\") on node \"crc\" DevicePath \"\"" Dec 04 10:32:37 crc kubenswrapper[4943]: I1204 10:32:37.688441 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8cdr\" (UniqueName: \"kubernetes.io/projected/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-kube-api-access-j8cdr\") on node \"crc\" DevicePath \"\"" Dec 04 10:32:37 crc kubenswrapper[4943]: I1204 10:32:37.688452 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:32:38 crc kubenswrapper[4943]: I1204 10:32:38.075280 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" event={"ID":"21b39ac1-11f1-4bf2-bdd8-6e93062ecafe","Type":"ContainerDied","Data":"610c50dacec3f35e6c65291a668605f4c6b3f3877a33cadb3bb4dd4e24246b59"} Dec 04 10:32:38 crc kubenswrapper[4943]: I1204 10:32:38.075334 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="610c50dacec3f35e6c65291a668605f4c6b3f3877a33cadb3bb4dd4e24246b59" Dec 04 10:32:38 crc kubenswrapper[4943]: I1204 10:32:38.075288 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.454934 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8"] Dec 04 10:32:42 crc kubenswrapper[4943]: E1204 10:32:42.456007 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" containerName="util" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.456037 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" containerName="util" Dec 04 10:32:42 crc kubenswrapper[4943]: E1204 10:32:42.456062 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" containerName="pull" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.456069 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" containerName="pull" Dec 04 10:32:42 crc kubenswrapper[4943]: E1204 10:32:42.456087 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" containerName="extract" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.456094 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" containerName="extract" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.456296 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" containerName="extract" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.457613 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.460289 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4xrhn" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.461094 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.471277 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8"] Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.552380 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-webhook-cert\") pod \"infra-operator-controller-manager-db7f8fc9c-qsqx8\" (UID: \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\") " pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.552439 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgd9s\" (UniqueName: \"kubernetes.io/projected/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-kube-api-access-dgd9s\") pod \"infra-operator-controller-manager-db7f8fc9c-qsqx8\" (UID: \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\") " pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.552483 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-apiservice-cert\") pod \"infra-operator-controller-manager-db7f8fc9c-qsqx8\" (UID: \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\") " pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.653973 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-webhook-cert\") pod \"infra-operator-controller-manager-db7f8fc9c-qsqx8\" (UID: \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\") " pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.654014 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgd9s\" (UniqueName: \"kubernetes.io/projected/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-kube-api-access-dgd9s\") pod \"infra-operator-controller-manager-db7f8fc9c-qsqx8\" (UID: \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\") " pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.654066 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-apiservice-cert\") pod \"infra-operator-controller-manager-db7f8fc9c-qsqx8\" (UID: \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\") " pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.661944 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-apiservice-cert\") pod \"infra-operator-controller-manager-db7f8fc9c-qsqx8\" (UID: \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\") " pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.673342 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgd9s\" (UniqueName: \"kubernetes.io/projected/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-kube-api-access-dgd9s\") pod \"infra-operator-controller-manager-db7f8fc9c-qsqx8\" (UID: \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\") " pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.673532 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-webhook-cert\") pod \"infra-operator-controller-manager-db7f8fc9c-qsqx8\" (UID: \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\") " pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.775422 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:42 crc kubenswrapper[4943]: I1204 10:32:42.989875 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8"] Dec 04 10:32:43 crc kubenswrapper[4943]: W1204 10:32:43.005841 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e7c05b7_ea2c_4771_ac23_f2217871d2b3.slice/crio-236d3612fb0c2f98b72196abd9ea9aa30cf4650f23fd66b90a25fa6acf698fe9 WatchSource:0}: Error finding container 236d3612fb0c2f98b72196abd9ea9aa30cf4650f23fd66b90a25fa6acf698fe9: Status 404 returned error can't find the container with id 236d3612fb0c2f98b72196abd9ea9aa30cf4650f23fd66b90a25fa6acf698fe9 Dec 04 10:32:43 crc kubenswrapper[4943]: I1204 10:32:43.104670 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" event={"ID":"5e7c05b7-ea2c-4771-ac23-f2217871d2b3","Type":"ContainerStarted","Data":"236d3612fb0c2f98b72196abd9ea9aa30cf4650f23fd66b90a25fa6acf698fe9"} Dec 04 10:32:45 crc kubenswrapper[4943]: I1204 10:32:45.130440 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" event={"ID":"5e7c05b7-ea2c-4771-ac23-f2217871d2b3","Type":"ContainerStarted","Data":"6442c1ee2235ed93b4bda925ce0d21931ebcb953ac8c2184b305eedc4dfe77a2"} Dec 04 10:32:47 crc kubenswrapper[4943]: I1204 10:32:47.148677 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" event={"ID":"5e7c05b7-ea2c-4771-ac23-f2217871d2b3","Type":"ContainerStarted","Data":"724e34ee83c66974dc2a6c83b4871b4fa49659cace9f6f61b91581129c990efb"} Dec 04 10:32:47 crc kubenswrapper[4943]: I1204 10:32:47.149069 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:47 crc kubenswrapper[4943]: I1204 10:32:47.168257 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" podStartSLOduration=1.385165308 podStartE2EDuration="5.168224084s" podCreationTimestamp="2025-12-04 10:32:42 +0000 UTC" firstStartedPulling="2025-12-04 10:32:43.008421152 +0000 UTC m=+1011.597397020" lastFinishedPulling="2025-12-04 10:32:46.791479928 +0000 UTC m=+1015.380455796" observedRunningTime="2025-12-04 10:32:47.165826027 +0000 UTC m=+1015.754801915" watchObservedRunningTime="2025-12-04 10:32:47.168224084 +0000 UTC m=+1015.757199952" Dec 04 10:32:48 crc kubenswrapper[4943]: I1204 10:32:48.629041 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:32:48 crc kubenswrapper[4943]: I1204 10:32:48.629113 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:32:48 crc kubenswrapper[4943]: I1204 10:32:48.629168 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:32:48 crc kubenswrapper[4943]: I1204 10:32:48.629847 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"41363b73659bf0d33a8ed7f51e273d64bed37bf7e075e5c6bc173eb617215207"} pod="openshift-machine-config-operator/machine-config-daemon-kswzd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 10:32:48 crc kubenswrapper[4943]: I1204 10:32:48.629922 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" containerID="cri-o://41363b73659bf0d33a8ed7f51e273d64bed37bf7e075e5c6bc173eb617215207" gracePeriod=600 Dec 04 10:32:49 crc kubenswrapper[4943]: I1204 10:32:49.178171 4943 generic.go:334] "Generic (PLEG): container finished" podID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerID="41363b73659bf0d33a8ed7f51e273d64bed37bf7e075e5c6bc173eb617215207" exitCode=0 Dec 04 10:32:49 crc kubenswrapper[4943]: I1204 10:32:49.178295 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerDied","Data":"41363b73659bf0d33a8ed7f51e273d64bed37bf7e075e5c6bc173eb617215207"} Dec 04 10:32:49 crc kubenswrapper[4943]: I1204 10:32:49.197472 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerStarted","Data":"39ef936964dcabd0fd14176e3d0e5709046494c4af42c583afefd4c04502ae7a"} Dec 04 10:32:49 crc kubenswrapper[4943]: I1204 10:32:49.197919 4943 scope.go:117] "RemoveContainer" containerID="2a882281b91801c61cd5c5ba5bc934cd8e35cb87485c0c878f8c0bf55706b5c5" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.341215 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.342563 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.344572 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openshift-service-ca.crt" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.344590 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openstack-config-data" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.345119 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openstack-scripts" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.346643 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"galera-openstack-dockercfg-qdzrv" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.347978 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"kube-root-ca.crt" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.360602 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.363789 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.365069 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.369184 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.370502 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.454585 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.461272 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.508690 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.508733 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-operator-scripts\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.508760 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrd6b\" (UniqueName: \"kubernetes.io/projected/55950271-9dac-4a16-8cca-5288ddb40dba-kube-api-access-vrd6b\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.508786 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e75cbefc-6761-45f0-86a7-d0ada42d35db-config-data-generated\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.508804 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdcfx\" (UniqueName: \"kubernetes.io/projected/90592895-1eb8-4b3b-bdb3-746b364256c1-kube-api-access-gdcfx\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.508824 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-kolla-config\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.508845 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/55950271-9dac-4a16-8cca-5288ddb40dba-config-data-generated\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.508867 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdwmr\" (UniqueName: \"kubernetes.io/projected/e75cbefc-6761-45f0-86a7-d0ada42d35db-kube-api-access-bdwmr\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.508894 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-operator-scripts\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.508987 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/90592895-1eb8-4b3b-bdb3-746b364256c1-config-data-generated\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.509028 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-kolla-config\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.509069 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.509101 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.509120 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-config-data-default\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.509143 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-config-data-default\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.509179 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-config-data-default\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.509227 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-kolla-config\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.509253 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-operator-scripts\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610258 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610308 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-config-data-default\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610334 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-config-data-default\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610360 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-config-data-default\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610385 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-kolla-config\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610401 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-operator-scripts\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610426 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610442 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-operator-scripts\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610458 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrd6b\" (UniqueName: \"kubernetes.io/projected/55950271-9dac-4a16-8cca-5288ddb40dba-kube-api-access-vrd6b\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610481 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e75cbefc-6761-45f0-86a7-d0ada42d35db-config-data-generated\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610495 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdcfx\" (UniqueName: \"kubernetes.io/projected/90592895-1eb8-4b3b-bdb3-746b364256c1-kube-api-access-gdcfx\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610513 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-kolla-config\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610533 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/55950271-9dac-4a16-8cca-5288ddb40dba-config-data-generated\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610551 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdwmr\" (UniqueName: \"kubernetes.io/projected/e75cbefc-6761-45f0-86a7-d0ada42d35db-kube-api-access-bdwmr\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610570 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-operator-scripts\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610587 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/90592895-1eb8-4b3b-bdb3-746b364256c1-config-data-generated\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610609 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-kolla-config\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610638 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610708 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") device mount path \"/mnt/openstack/pv01\"" pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.610782 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") device mount path \"/mnt/openstack/pv06\"" pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.611102 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-kolla-config\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.611330 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-config-data-default\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.611720 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-config-data-default\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.611901 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-config-data-default\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.612034 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") device mount path \"/mnt/openstack/pv03\"" pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.612289 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e75cbefc-6761-45f0-86a7-d0ada42d35db-config-data-generated\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.613387 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-kolla-config\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.613409 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-operator-scripts\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.613705 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-operator-scripts\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.613737 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/90592895-1eb8-4b3b-bdb3-746b364256c1-config-data-generated\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.613867 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/55950271-9dac-4a16-8cca-5288ddb40dba-config-data-generated\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.615614 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-operator-scripts\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.617529 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-kolla-config\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.633490 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdcfx\" (UniqueName: \"kubernetes.io/projected/90592895-1eb8-4b3b-bdb3-746b364256c1-kube-api-access-gdcfx\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.635445 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.637932 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrd6b\" (UniqueName: \"kubernetes.io/projected/55950271-9dac-4a16-8cca-5288ddb40dba-kube-api-access-vrd6b\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.639359 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdwmr\" (UniqueName: \"kubernetes.io/projected/e75cbefc-6761-45f0-86a7-d0ada42d35db-kube-api-access-bdwmr\") pod \"openstack-galera-1\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.641712 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-2\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.642238 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.723371 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.737142 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:32:51 crc kubenswrapper[4943]: I1204 10:32:51.760807 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:32:52 crc kubenswrapper[4943]: I1204 10:32:52.201033 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 04 10:32:52 crc kubenswrapper[4943]: I1204 10:32:52.208348 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 04 10:32:52 crc kubenswrapper[4943]: W1204 10:32:52.210634 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode75cbefc_6761_45f0_86a7_d0ada42d35db.slice/crio-5569a85e24f4743cecf769628abfb8142d82ce9c24a58caadcae27b1f52d1a6a WatchSource:0}: Error finding container 5569a85e24f4743cecf769628abfb8142d82ce9c24a58caadcae27b1f52d1a6a: Status 404 returned error can't find the container with id 5569a85e24f4743cecf769628abfb8142d82ce9c24a58caadcae27b1f52d1a6a Dec 04 10:32:52 crc kubenswrapper[4943]: W1204 10:32:52.211683 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90592895_1eb8_4b3b_bdb3_746b364256c1.slice/crio-2247fd983dab29b96e207181e6e8d9e928e9d84210cbd750b079d9b503a9bb7d WatchSource:0}: Error finding container 2247fd983dab29b96e207181e6e8d9e928e9d84210cbd750b079d9b503a9bb7d: Status 404 returned error can't find the container with id 2247fd983dab29b96e207181e6e8d9e928e9d84210cbd750b079d9b503a9bb7d Dec 04 10:32:52 crc kubenswrapper[4943]: I1204 10:32:52.255546 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 04 10:32:52 crc kubenswrapper[4943]: W1204 10:32:52.259668 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55950271_9dac_4a16_8cca_5288ddb40dba.slice/crio-52c85e45d636b61dce12cef4dcf398c3860441e0964b2dcfeb2d06a0bdd1ce70 WatchSource:0}: Error finding container 52c85e45d636b61dce12cef4dcf398c3860441e0964b2dcfeb2d06a0bdd1ce70: Status 404 returned error can't find the container with id 52c85e45d636b61dce12cef4dcf398c3860441e0964b2dcfeb2d06a0bdd1ce70 Dec 04 10:32:52 crc kubenswrapper[4943]: I1204 10:32:52.781025 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:32:53 crc kubenswrapper[4943]: I1204 10:32:53.228585 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"90592895-1eb8-4b3b-bdb3-746b364256c1","Type":"ContainerStarted","Data":"2247fd983dab29b96e207181e6e8d9e928e9d84210cbd750b079d9b503a9bb7d"} Dec 04 10:32:53 crc kubenswrapper[4943]: I1204 10:32:53.232360 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"e75cbefc-6761-45f0-86a7-d0ada42d35db","Type":"ContainerStarted","Data":"5569a85e24f4743cecf769628abfb8142d82ce9c24a58caadcae27b1f52d1a6a"} Dec 04 10:32:53 crc kubenswrapper[4943]: I1204 10:32:53.234068 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"55950271-9dac-4a16-8cca-5288ddb40dba","Type":"ContainerStarted","Data":"52c85e45d636b61dce12cef4dcf398c3860441e0964b2dcfeb2d06a0bdd1ce70"} Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.018508 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.020392 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.027702 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"memcached-memcached-dockercfg-chlhk" Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.028265 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"memcached-config-data" Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.095073 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/197663a6-bdeb-49ee-ad21-9b76212842e9-kolla-config\") pod \"memcached-0\" (UID: \"197663a6-bdeb-49ee-ad21-9b76212842e9\") " pod="cinder-kuttl-tests/memcached-0" Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.095160 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qdzx\" (UniqueName: \"kubernetes.io/projected/197663a6-bdeb-49ee-ad21-9b76212842e9-kube-api-access-8qdzx\") pod \"memcached-0\" (UID: \"197663a6-bdeb-49ee-ad21-9b76212842e9\") " pod="cinder-kuttl-tests/memcached-0" Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.095190 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/197663a6-bdeb-49ee-ad21-9b76212842e9-config-data\") pod \"memcached-0\" (UID: \"197663a6-bdeb-49ee-ad21-9b76212842e9\") " pod="cinder-kuttl-tests/memcached-0" Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.109659 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.196238 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/197663a6-bdeb-49ee-ad21-9b76212842e9-kolla-config\") pod \"memcached-0\" (UID: \"197663a6-bdeb-49ee-ad21-9b76212842e9\") " pod="cinder-kuttl-tests/memcached-0" Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.196321 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qdzx\" (UniqueName: \"kubernetes.io/projected/197663a6-bdeb-49ee-ad21-9b76212842e9-kube-api-access-8qdzx\") pod \"memcached-0\" (UID: \"197663a6-bdeb-49ee-ad21-9b76212842e9\") " pod="cinder-kuttl-tests/memcached-0" Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.196349 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/197663a6-bdeb-49ee-ad21-9b76212842e9-config-data\") pod \"memcached-0\" (UID: \"197663a6-bdeb-49ee-ad21-9b76212842e9\") " pod="cinder-kuttl-tests/memcached-0" Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.197008 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/197663a6-bdeb-49ee-ad21-9b76212842e9-kolla-config\") pod \"memcached-0\" (UID: \"197663a6-bdeb-49ee-ad21-9b76212842e9\") " pod="cinder-kuttl-tests/memcached-0" Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.197044 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/197663a6-bdeb-49ee-ad21-9b76212842e9-config-data\") pod \"memcached-0\" (UID: \"197663a6-bdeb-49ee-ad21-9b76212842e9\") " pod="cinder-kuttl-tests/memcached-0" Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.237302 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qdzx\" (UniqueName: \"kubernetes.io/projected/197663a6-bdeb-49ee-ad21-9b76212842e9-kube-api-access-8qdzx\") pod \"memcached-0\" (UID: \"197663a6-bdeb-49ee-ad21-9b76212842e9\") " pod="cinder-kuttl-tests/memcached-0" Dec 04 10:32:58 crc kubenswrapper[4943]: I1204 10:32:58.340093 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Dec 04 10:33:00 crc kubenswrapper[4943]: I1204 10:33:00.739285 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-s99cd"] Dec 04 10:33:00 crc kubenswrapper[4943]: I1204 10:33:00.740511 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" Dec 04 10:33:00 crc kubenswrapper[4943]: I1204 10:33:00.742814 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-c5c6h" Dec 04 10:33:00 crc kubenswrapper[4943]: I1204 10:33:00.754052 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-s99cd"] Dec 04 10:33:00 crc kubenswrapper[4943]: I1204 10:33:00.832682 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvph2\" (UniqueName: \"kubernetes.io/projected/34001aba-ee9d-4c56-9895-009f4fea5926-kube-api-access-kvph2\") pod \"rabbitmq-cluster-operator-index-s99cd\" (UID: \"34001aba-ee9d-4c56-9895-009f4fea5926\") " pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" Dec 04 10:33:00 crc kubenswrapper[4943]: I1204 10:33:00.934400 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvph2\" (UniqueName: \"kubernetes.io/projected/34001aba-ee9d-4c56-9895-009f4fea5926-kube-api-access-kvph2\") pod \"rabbitmq-cluster-operator-index-s99cd\" (UID: \"34001aba-ee9d-4c56-9895-009f4fea5926\") " pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" Dec 04 10:33:00 crc kubenswrapper[4943]: I1204 10:33:00.955449 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvph2\" (UniqueName: \"kubernetes.io/projected/34001aba-ee9d-4c56-9895-009f4fea5926-kube-api-access-kvph2\") pod \"rabbitmq-cluster-operator-index-s99cd\" (UID: \"34001aba-ee9d-4c56-9895-009f4fea5926\") " pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" Dec 04 10:33:01 crc kubenswrapper[4943]: I1204 10:33:01.070942 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" Dec 04 10:33:01 crc kubenswrapper[4943]: I1204 10:33:01.583346 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-s99cd"] Dec 04 10:33:01 crc kubenswrapper[4943]: I1204 10:33:01.674107 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 04 10:33:01 crc kubenswrapper[4943]: W1204 10:33:01.680721 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod197663a6_bdeb_49ee_ad21_9b76212842e9.slice/crio-511d05696b35e326de2476ac9349af3bde12c28d35da0deeabce0f357e038054 WatchSource:0}: Error finding container 511d05696b35e326de2476ac9349af3bde12c28d35da0deeabce0f357e038054: Status 404 returned error can't find the container with id 511d05696b35e326de2476ac9349af3bde12c28d35da0deeabce0f357e038054 Dec 04 10:33:02 crc kubenswrapper[4943]: I1204 10:33:02.309826 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" event={"ID":"34001aba-ee9d-4c56-9895-009f4fea5926","Type":"ContainerStarted","Data":"ff3ec7c7a11cc2e2d43ddddf298724e44b197d6ceb8d3016c2b54bce5178e49d"} Dec 04 10:33:02 crc kubenswrapper[4943]: I1204 10:33:02.311930 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"e75cbefc-6761-45f0-86a7-d0ada42d35db","Type":"ContainerStarted","Data":"0e58835abd4cbec14ccbc6596bb0acee42ba89f53d70bca8690701acfd3aee55"} Dec 04 10:33:02 crc kubenswrapper[4943]: I1204 10:33:02.316929 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"55950271-9dac-4a16-8cca-5288ddb40dba","Type":"ContainerStarted","Data":"3e2a8a68745c611b3c975bedaa991c7e4881c6bcb3a24bf94d91e2ba5b91a7e4"} Dec 04 10:33:02 crc kubenswrapper[4943]: I1204 10:33:02.319075 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"197663a6-bdeb-49ee-ad21-9b76212842e9","Type":"ContainerStarted","Data":"511d05696b35e326de2476ac9349af3bde12c28d35da0deeabce0f357e038054"} Dec 04 10:33:02 crc kubenswrapper[4943]: I1204 10:33:02.321126 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"90592895-1eb8-4b3b-bdb3-746b364256c1","Type":"ContainerStarted","Data":"b5258258825b305b21c8434353b1f1e0f2f8aa633a7cd6542d026ebcfcf7643a"} Dec 04 10:33:06 crc kubenswrapper[4943]: I1204 10:33:06.354697 4943 generic.go:334] "Generic (PLEG): container finished" podID="90592895-1eb8-4b3b-bdb3-746b364256c1" containerID="b5258258825b305b21c8434353b1f1e0f2f8aa633a7cd6542d026ebcfcf7643a" exitCode=0 Dec 04 10:33:06 crc kubenswrapper[4943]: I1204 10:33:06.354799 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"90592895-1eb8-4b3b-bdb3-746b364256c1","Type":"ContainerDied","Data":"b5258258825b305b21c8434353b1f1e0f2f8aa633a7cd6542d026ebcfcf7643a"} Dec 04 10:33:06 crc kubenswrapper[4943]: I1204 10:33:06.357654 4943 generic.go:334] "Generic (PLEG): container finished" podID="e75cbefc-6761-45f0-86a7-d0ada42d35db" containerID="0e58835abd4cbec14ccbc6596bb0acee42ba89f53d70bca8690701acfd3aee55" exitCode=0 Dec 04 10:33:06 crc kubenswrapper[4943]: I1204 10:33:06.357707 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"e75cbefc-6761-45f0-86a7-d0ada42d35db","Type":"ContainerDied","Data":"0e58835abd4cbec14ccbc6596bb0acee42ba89f53d70bca8690701acfd3aee55"} Dec 04 10:33:06 crc kubenswrapper[4943]: I1204 10:33:06.360371 4943 generic.go:334] "Generic (PLEG): container finished" podID="55950271-9dac-4a16-8cca-5288ddb40dba" containerID="3e2a8a68745c611b3c975bedaa991c7e4881c6bcb3a24bf94d91e2ba5b91a7e4" exitCode=0 Dec 04 10:33:06 crc kubenswrapper[4943]: I1204 10:33:06.360403 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"55950271-9dac-4a16-8cca-5288ddb40dba","Type":"ContainerDied","Data":"3e2a8a68745c611b3c975bedaa991c7e4881c6bcb3a24bf94d91e2ba5b91a7e4"} Dec 04 10:33:09 crc kubenswrapper[4943]: I1204 10:33:09.395950 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"55950271-9dac-4a16-8cca-5288ddb40dba","Type":"ContainerStarted","Data":"9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc"} Dec 04 10:33:09 crc kubenswrapper[4943]: I1204 10:33:09.399170 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"90592895-1eb8-4b3b-bdb3-746b364256c1","Type":"ContainerStarted","Data":"f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286"} Dec 04 10:33:09 crc kubenswrapper[4943]: I1204 10:33:09.401166 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"e75cbefc-6761-45f0-86a7-d0ada42d35db","Type":"ContainerStarted","Data":"61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9"} Dec 04 10:33:09 crc kubenswrapper[4943]: I1204 10:33:09.414463 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-0" podStartSLOduration=10.361390278 podStartE2EDuration="19.414447684s" podCreationTimestamp="2025-12-04 10:32:50 +0000 UTC" firstStartedPulling="2025-12-04 10:32:52.262481195 +0000 UTC m=+1020.851457063" lastFinishedPulling="2025-12-04 10:33:01.315538601 +0000 UTC m=+1029.904514469" observedRunningTime="2025-12-04 10:33:09.413340821 +0000 UTC m=+1038.002316689" watchObservedRunningTime="2025-12-04 10:33:09.414447684 +0000 UTC m=+1038.003423552" Dec 04 10:33:09 crc kubenswrapper[4943]: I1204 10:33:09.434001 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-1" podStartSLOduration=10.368692408 podStartE2EDuration="19.433981248s" podCreationTimestamp="2025-12-04 10:32:50 +0000 UTC" firstStartedPulling="2025-12-04 10:32:52.212765767 +0000 UTC m=+1020.801741625" lastFinishedPulling="2025-12-04 10:33:01.278054597 +0000 UTC m=+1029.867030465" observedRunningTime="2025-12-04 10:33:09.430794606 +0000 UTC m=+1038.019770484" watchObservedRunningTime="2025-12-04 10:33:09.433981248 +0000 UTC m=+1038.022957136" Dec 04 10:33:09 crc kubenswrapper[4943]: I1204 10:33:09.454049 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-2" podStartSLOduration=10.391973888 podStartE2EDuration="19.454026917s" podCreationTimestamp="2025-12-04 10:32:50 +0000 UTC" firstStartedPulling="2025-12-04 10:32:52.213849328 +0000 UTC m=+1020.802825196" lastFinishedPulling="2025-12-04 10:33:01.275902357 +0000 UTC m=+1029.864878225" observedRunningTime="2025-12-04 10:33:09.448552028 +0000 UTC m=+1038.037527916" watchObservedRunningTime="2025-12-04 10:33:09.454026917 +0000 UTC m=+1038.043002785" Dec 04 10:33:10 crc kubenswrapper[4943]: I1204 10:33:10.408567 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" event={"ID":"34001aba-ee9d-4c56-9895-009f4fea5926","Type":"ContainerStarted","Data":"2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59"} Dec 04 10:33:10 crc kubenswrapper[4943]: I1204 10:33:10.411839 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"197663a6-bdeb-49ee-ad21-9b76212842e9","Type":"ContainerStarted","Data":"982cf322919a66576035509dc13c52ef8d2f4017b05fe27ba6629d63ee3e4f1a"} Dec 04 10:33:10 crc kubenswrapper[4943]: I1204 10:33:10.412083 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/memcached-0" Dec 04 10:33:10 crc kubenswrapper[4943]: I1204 10:33:10.425448 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" podStartSLOduration=2.461659158 podStartE2EDuration="10.42542568s" podCreationTimestamp="2025-12-04 10:33:00 +0000 UTC" firstStartedPulling="2025-12-04 10:33:01.587299367 +0000 UTC m=+1030.176275235" lastFinishedPulling="2025-12-04 10:33:09.551065889 +0000 UTC m=+1038.140041757" observedRunningTime="2025-12-04 10:33:10.424316957 +0000 UTC m=+1039.013292815" watchObservedRunningTime="2025-12-04 10:33:10.42542568 +0000 UTC m=+1039.014401548" Dec 04 10:33:10 crc kubenswrapper[4943]: I1204 10:33:10.441758 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/memcached-0" podStartSLOduration=6.957650238 podStartE2EDuration="13.44173434s" podCreationTimestamp="2025-12-04 10:32:57 +0000 UTC" firstStartedPulling="2025-12-04 10:33:01.696004832 +0000 UTC m=+1030.284980700" lastFinishedPulling="2025-12-04 10:33:08.180088934 +0000 UTC m=+1036.769064802" observedRunningTime="2025-12-04 10:33:10.438469396 +0000 UTC m=+1039.027445264" watchObservedRunningTime="2025-12-04 10:33:10.44173434 +0000 UTC m=+1039.030710208" Dec 04 10:33:11 crc kubenswrapper[4943]: I1204 10:33:11.071473 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" Dec 04 10:33:11 crc kubenswrapper[4943]: I1204 10:33:11.071532 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" Dec 04 10:33:11 crc kubenswrapper[4943]: I1204 10:33:11.094425 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" Dec 04 10:33:11 crc kubenswrapper[4943]: I1204 10:33:11.723924 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:33:11 crc kubenswrapper[4943]: I1204 10:33:11.724028 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:33:11 crc kubenswrapper[4943]: I1204 10:33:11.738463 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:33:11 crc kubenswrapper[4943]: I1204 10:33:11.738525 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:33:11 crc kubenswrapper[4943]: I1204 10:33:11.761885 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:33:11 crc kubenswrapper[4943]: I1204 10:33:11.761939 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:33:15 crc kubenswrapper[4943]: E1204 10:33:15.192155 4943 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.150:57192->38.102.83.150:36001: write tcp 38.102.83.150:57192->38.102.83.150:36001: write: connection reset by peer Dec 04 10:33:17 crc kubenswrapper[4943]: I1204 10:33:17.834408 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:33:17 crc kubenswrapper[4943]: I1204 10:33:17.905766 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:33:18 crc kubenswrapper[4943]: I1204 10:33:18.342617 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/memcached-0" Dec 04 10:33:21 crc kubenswrapper[4943]: I1204 10:33:21.096273 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" Dec 04 10:33:21 crc kubenswrapper[4943]: I1204 10:33:21.823989 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/openstack-galera-2" podUID="90592895-1eb8-4b3b-bdb3-746b364256c1" containerName="galera" probeResult="failure" output=< Dec 04 10:33:21 crc kubenswrapper[4943]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Dec 04 10:33:21 crc kubenswrapper[4943]: > Dec 04 10:33:27 crc kubenswrapper[4943]: I1204 10:33:27.992484 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk"] Dec 04 10:33:27 crc kubenswrapper[4943]: I1204 10:33:27.994802 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:27 crc kubenswrapper[4943]: I1204 10:33:27.998378 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6gzwg" Dec 04 10:33:28 crc kubenswrapper[4943]: I1204 10:33:28.004453 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk"] Dec 04 10:33:28 crc kubenswrapper[4943]: I1204 10:33:28.120142 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e9104f1-6500-4ca1-88a4-b7b912c51caa-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk\" (UID: \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:28 crc kubenswrapper[4943]: I1204 10:33:28.120352 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw7dm\" (UniqueName: \"kubernetes.io/projected/7e9104f1-6500-4ca1-88a4-b7b912c51caa-kube-api-access-gw7dm\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk\" (UID: \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:28 crc kubenswrapper[4943]: I1204 10:33:28.120454 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e9104f1-6500-4ca1-88a4-b7b912c51caa-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk\" (UID: \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:28 crc kubenswrapper[4943]: I1204 10:33:28.221892 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e9104f1-6500-4ca1-88a4-b7b912c51caa-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk\" (UID: \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:28 crc kubenswrapper[4943]: I1204 10:33:28.221943 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw7dm\" (UniqueName: \"kubernetes.io/projected/7e9104f1-6500-4ca1-88a4-b7b912c51caa-kube-api-access-gw7dm\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk\" (UID: \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:28 crc kubenswrapper[4943]: I1204 10:33:28.221986 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e9104f1-6500-4ca1-88a4-b7b912c51caa-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk\" (UID: \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:28 crc kubenswrapper[4943]: I1204 10:33:28.222448 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e9104f1-6500-4ca1-88a4-b7b912c51caa-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk\" (UID: \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:28 crc kubenswrapper[4943]: I1204 10:33:28.222680 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e9104f1-6500-4ca1-88a4-b7b912c51caa-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk\" (UID: \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:28 crc kubenswrapper[4943]: I1204 10:33:28.241590 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw7dm\" (UniqueName: \"kubernetes.io/projected/7e9104f1-6500-4ca1-88a4-b7b912c51caa-kube-api-access-gw7dm\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk\" (UID: \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:28 crc kubenswrapper[4943]: I1204 10:33:28.313814 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:28 crc kubenswrapper[4943]: I1204 10:33:28.787978 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk"] Dec 04 10:33:28 crc kubenswrapper[4943]: W1204 10:33:28.799886 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e9104f1_6500_4ca1_88a4_b7b912c51caa.slice/crio-a2bf5cbdb1d51db974ce4be306f3a3a3373b09b098b4d0f224e016a7a793b5d4 WatchSource:0}: Error finding container a2bf5cbdb1d51db974ce4be306f3a3a3373b09b098b4d0f224e016a7a793b5d4: Status 404 returned error can't find the container with id a2bf5cbdb1d51db974ce4be306f3a3a3373b09b098b4d0f224e016a7a793b5d4 Dec 04 10:33:29 crc kubenswrapper[4943]: I1204 10:33:29.540258 4943 generic.go:334] "Generic (PLEG): container finished" podID="7e9104f1-6500-4ca1-88a4-b7b912c51caa" containerID="258b8d02268fdb04b1901f1cdb2afb5e9e91515232a922f604a70c11976b76bc" exitCode=0 Dec 04 10:33:29 crc kubenswrapper[4943]: I1204 10:33:29.540360 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" event={"ID":"7e9104f1-6500-4ca1-88a4-b7b912c51caa","Type":"ContainerDied","Data":"258b8d02268fdb04b1901f1cdb2afb5e9e91515232a922f604a70c11976b76bc"} Dec 04 10:33:29 crc kubenswrapper[4943]: I1204 10:33:29.540579 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" event={"ID":"7e9104f1-6500-4ca1-88a4-b7b912c51caa","Type":"ContainerStarted","Data":"a2bf5cbdb1d51db974ce4be306f3a3a3373b09b098b4d0f224e016a7a793b5d4"} Dec 04 10:33:31 crc kubenswrapper[4943]: I1204 10:33:31.555227 4943 generic.go:334] "Generic (PLEG): container finished" podID="7e9104f1-6500-4ca1-88a4-b7b912c51caa" containerID="3ea468f3c2ce28512325954b214388854cdae92e4b9a990e314f27613d83be1f" exitCode=0 Dec 04 10:33:31 crc kubenswrapper[4943]: I1204 10:33:31.555305 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" event={"ID":"7e9104f1-6500-4ca1-88a4-b7b912c51caa","Type":"ContainerDied","Data":"3ea468f3c2ce28512325954b214388854cdae92e4b9a990e314f27613d83be1f"} Dec 04 10:33:31 crc kubenswrapper[4943]: I1204 10:33:31.664565 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:33:31 crc kubenswrapper[4943]: I1204 10:33:31.725864 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:33:32 crc kubenswrapper[4943]: I1204 10:33:32.565666 4943 generic.go:334] "Generic (PLEG): container finished" podID="7e9104f1-6500-4ca1-88a4-b7b912c51caa" containerID="d6149d6d408f5420971882aac7f59dd83734dcb21f29c76b5b83e3ab43a23f6e" exitCode=0 Dec 04 10:33:32 crc kubenswrapper[4943]: I1204 10:33:32.572800 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" event={"ID":"7e9104f1-6500-4ca1-88a4-b7b912c51caa","Type":"ContainerDied","Data":"d6149d6d408f5420971882aac7f59dd83734dcb21f29c76b5b83e3ab43a23f6e"} Dec 04 10:33:33 crc kubenswrapper[4943]: I1204 10:33:33.815069 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:33 crc kubenswrapper[4943]: I1204 10:33:33.923357 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw7dm\" (UniqueName: \"kubernetes.io/projected/7e9104f1-6500-4ca1-88a4-b7b912c51caa-kube-api-access-gw7dm\") pod \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\" (UID: \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\") " Dec 04 10:33:33 crc kubenswrapper[4943]: I1204 10:33:33.923428 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e9104f1-6500-4ca1-88a4-b7b912c51caa-bundle\") pod \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\" (UID: \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\") " Dec 04 10:33:33 crc kubenswrapper[4943]: I1204 10:33:33.923470 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e9104f1-6500-4ca1-88a4-b7b912c51caa-util\") pod \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\" (UID: \"7e9104f1-6500-4ca1-88a4-b7b912c51caa\") " Dec 04 10:33:33 crc kubenswrapper[4943]: I1204 10:33:33.924095 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e9104f1-6500-4ca1-88a4-b7b912c51caa-bundle" (OuterVolumeSpecName: "bundle") pod "7e9104f1-6500-4ca1-88a4-b7b912c51caa" (UID: "7e9104f1-6500-4ca1-88a4-b7b912c51caa"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:33:33 crc kubenswrapper[4943]: I1204 10:33:33.936142 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9104f1-6500-4ca1-88a4-b7b912c51caa-kube-api-access-gw7dm" (OuterVolumeSpecName: "kube-api-access-gw7dm") pod "7e9104f1-6500-4ca1-88a4-b7b912c51caa" (UID: "7e9104f1-6500-4ca1-88a4-b7b912c51caa"). InnerVolumeSpecName "kube-api-access-gw7dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:33:33 crc kubenswrapper[4943]: I1204 10:33:33.938059 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e9104f1-6500-4ca1-88a4-b7b912c51caa-util" (OuterVolumeSpecName: "util") pod "7e9104f1-6500-4ca1-88a4-b7b912c51caa" (UID: "7e9104f1-6500-4ca1-88a4-b7b912c51caa"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:33:34 crc kubenswrapper[4943]: I1204 10:33:34.025156 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e9104f1-6500-4ca1-88a4-b7b912c51caa-util\") on node \"crc\" DevicePath \"\"" Dec 04 10:33:34 crc kubenswrapper[4943]: I1204 10:33:34.025212 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw7dm\" (UniqueName: \"kubernetes.io/projected/7e9104f1-6500-4ca1-88a4-b7b912c51caa-kube-api-access-gw7dm\") on node \"crc\" DevicePath \"\"" Dec 04 10:33:34 crc kubenswrapper[4943]: I1204 10:33:34.025223 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e9104f1-6500-4ca1-88a4-b7b912c51caa-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:33:34 crc kubenswrapper[4943]: I1204 10:33:34.581369 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" event={"ID":"7e9104f1-6500-4ca1-88a4-b7b912c51caa","Type":"ContainerDied","Data":"a2bf5cbdb1d51db974ce4be306f3a3a3373b09b098b4d0f224e016a7a793b5d4"} Dec 04 10:33:34 crc kubenswrapper[4943]: I1204 10:33:34.581426 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2bf5cbdb1d51db974ce4be306f3a3a3373b09b098b4d0f224e016a7a793b5d4" Dec 04 10:33:34 crc kubenswrapper[4943]: I1204 10:33:34.581518 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk" Dec 04 10:33:34 crc kubenswrapper[4943]: I1204 10:33:34.737470 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:33:34 crc kubenswrapper[4943]: I1204 10:33:34.797810 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.042888 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97"] Dec 04 10:33:43 crc kubenswrapper[4943]: E1204 10:33:43.043485 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9104f1-6500-4ca1-88a4-b7b912c51caa" containerName="util" Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.043500 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9104f1-6500-4ca1-88a4-b7b912c51caa" containerName="util" Dec 04 10:33:43 crc kubenswrapper[4943]: E1204 10:33:43.043527 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9104f1-6500-4ca1-88a4-b7b912c51caa" containerName="pull" Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.043534 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9104f1-6500-4ca1-88a4-b7b912c51caa" containerName="pull" Dec 04 10:33:43 crc kubenswrapper[4943]: E1204 10:33:43.043544 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9104f1-6500-4ca1-88a4-b7b912c51caa" containerName="extract" Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.043552 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9104f1-6500-4ca1-88a4-b7b912c51caa" containerName="extract" Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.043692 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9104f1-6500-4ca1-88a4-b7b912c51caa" containerName="extract" Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.044216 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.046633 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-48lrr" Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.054424 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97"] Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.197452 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c98l\" (UniqueName: \"kubernetes.io/projected/9cbafb93-24ac-4d39-bd39-0869654fb225-kube-api-access-7c98l\") pod \"rabbitmq-cluster-operator-779fc9694b-gsn97\" (UID: \"9cbafb93-24ac-4d39-bd39-0869654fb225\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.299578 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c98l\" (UniqueName: \"kubernetes.io/projected/9cbafb93-24ac-4d39-bd39-0869654fb225-kube-api-access-7c98l\") pod \"rabbitmq-cluster-operator-779fc9694b-gsn97\" (UID: \"9cbafb93-24ac-4d39-bd39-0869654fb225\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.332516 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c98l\" (UniqueName: \"kubernetes.io/projected/9cbafb93-24ac-4d39-bd39-0869654fb225-kube-api-access-7c98l\") pod \"rabbitmq-cluster-operator-779fc9694b-gsn97\" (UID: \"9cbafb93-24ac-4d39-bd39-0869654fb225\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.370016 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.573479 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97"] Dec 04 10:33:43 crc kubenswrapper[4943]: I1204 10:33:43.638969 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" event={"ID":"9cbafb93-24ac-4d39-bd39-0869654fb225","Type":"ContainerStarted","Data":"3f6066fb3bd1d4fce48e876f041587fd7a498ca0ebc986b96a0dc5dd74acadb2"} Dec 04 10:33:48 crc kubenswrapper[4943]: I1204 10:33:48.700617 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" event={"ID":"9cbafb93-24ac-4d39-bd39-0869654fb225","Type":"ContainerStarted","Data":"f42b9950e81784930b83aced47ed9ea7c6642570c940c3d6eb86bb6ce8ab6115"} Dec 04 10:33:48 crc kubenswrapper[4943]: I1204 10:33:48.723302 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" podStartSLOduration=1.621551228 podStartE2EDuration="5.723276343s" podCreationTimestamp="2025-12-04 10:33:43 +0000 UTC" firstStartedPulling="2025-12-04 10:33:43.601574042 +0000 UTC m=+1072.190549920" lastFinishedPulling="2025-12-04 10:33:47.703299167 +0000 UTC m=+1076.292275035" observedRunningTime="2025-12-04 10:33:48.720996557 +0000 UTC m=+1077.309972425" watchObservedRunningTime="2025-12-04 10:33:48.723276343 +0000 UTC m=+1077.312252231" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.088312 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.089860 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.091869 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-server-dockercfg-v5w8x" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.092647 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-default-user" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.103478 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"rabbitmq-server-conf" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.103478 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-erlang-cookie" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.108166 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"rabbitmq-plugins-conf" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.140635 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.208065 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.208147 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/057bf926-e2de-427f-9288-345fa9a798a6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.208246 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.208324 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.208369 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/057bf926-e2de-427f-9288-345fa9a798a6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.208417 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpx9s\" (UniqueName: \"kubernetes.io/projected/057bf926-e2de-427f-9288-345fa9a798a6-kube-api-access-tpx9s\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.208442 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7afe53b7-993a-480f-963a-183d18cbdf12\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7afe53b7-993a-480f-963a-183d18cbdf12\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.208468 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/057bf926-e2de-427f-9288-345fa9a798a6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.310248 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.310317 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/057bf926-e2de-427f-9288-345fa9a798a6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.310340 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.310357 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.310388 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/057bf926-e2de-427f-9288-345fa9a798a6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.310417 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpx9s\" (UniqueName: \"kubernetes.io/projected/057bf926-e2de-427f-9288-345fa9a798a6-kube-api-access-tpx9s\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.310445 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7afe53b7-993a-480f-963a-183d18cbdf12\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7afe53b7-993a-480f-963a-183d18cbdf12\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.310475 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/057bf926-e2de-427f-9288-345fa9a798a6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.310817 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.311338 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/057bf926-e2de-427f-9288-345fa9a798a6-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.312004 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.313639 4943 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.313673 4943 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7afe53b7-993a-480f-963a-183d18cbdf12\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7afe53b7-993a-480f-963a-183d18cbdf12\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/39da8ad30f8651e87617b3941ea4712e0d75881a143db5efa7cb19646ae7002e/globalmount\"" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.317750 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/057bf926-e2de-427f-9288-345fa9a798a6-pod-info\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.317946 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/057bf926-e2de-427f-9288-345fa9a798a6-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.317950 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.328590 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpx9s\" (UniqueName: \"kubernetes.io/projected/057bf926-e2de-427f-9288-345fa9a798a6-kube-api-access-tpx9s\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.341820 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7afe53b7-993a-480f-963a-183d18cbdf12\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7afe53b7-993a-480f-963a-183d18cbdf12\") pod \"rabbitmq-server-0\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.414004 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:33:51 crc kubenswrapper[4943]: I1204 10:33:51.846105 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 04 10:33:52 crc kubenswrapper[4943]: I1204 10:33:52.723534 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"057bf926-e2de-427f-9288-345fa9a798a6","Type":"ContainerStarted","Data":"874b07b83b2f9bd524cdaee4fb66fae48a808c137190f2d0fb3011fdbbc911db"} Dec 04 10:33:52 crc kubenswrapper[4943]: I1204 10:33:52.736330 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-p744p"] Dec 04 10:33:52 crc kubenswrapper[4943]: I1204 10:33:52.737341 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-p744p" Dec 04 10:33:52 crc kubenswrapper[4943]: I1204 10:33:52.739421 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-c4vtn" Dec 04 10:33:52 crc kubenswrapper[4943]: I1204 10:33:52.744893 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-p744p"] Dec 04 10:33:52 crc kubenswrapper[4943]: I1204 10:33:52.859728 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6vz9\" (UniqueName: \"kubernetes.io/projected/f5acada6-99e9-4109-b6fe-0b636e85a756-kube-api-access-p6vz9\") pod \"keystone-operator-index-p744p\" (UID: \"f5acada6-99e9-4109-b6fe-0b636e85a756\") " pod="openstack-operators/keystone-operator-index-p744p" Dec 04 10:33:52 crc kubenswrapper[4943]: I1204 10:33:52.962187 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6vz9\" (UniqueName: \"kubernetes.io/projected/f5acada6-99e9-4109-b6fe-0b636e85a756-kube-api-access-p6vz9\") pod \"keystone-operator-index-p744p\" (UID: \"f5acada6-99e9-4109-b6fe-0b636e85a756\") " pod="openstack-operators/keystone-operator-index-p744p" Dec 04 10:33:52 crc kubenswrapper[4943]: I1204 10:33:52.982069 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6vz9\" (UniqueName: \"kubernetes.io/projected/f5acada6-99e9-4109-b6fe-0b636e85a756-kube-api-access-p6vz9\") pod \"keystone-operator-index-p744p\" (UID: \"f5acada6-99e9-4109-b6fe-0b636e85a756\") " pod="openstack-operators/keystone-operator-index-p744p" Dec 04 10:33:53 crc kubenswrapper[4943]: I1204 10:33:53.063352 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-p744p" Dec 04 10:33:53 crc kubenswrapper[4943]: I1204 10:33:53.303018 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-p744p"] Dec 04 10:33:53 crc kubenswrapper[4943]: W1204 10:33:53.313085 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5acada6_99e9_4109_b6fe_0b636e85a756.slice/crio-97227c3e16bdc36f994f5f805ddf9fd10fb69469b9539838c257329919d2317a WatchSource:0}: Error finding container 97227c3e16bdc36f994f5f805ddf9fd10fb69469b9539838c257329919d2317a: Status 404 returned error can't find the container with id 97227c3e16bdc36f994f5f805ddf9fd10fb69469b9539838c257329919d2317a Dec 04 10:33:53 crc kubenswrapper[4943]: I1204 10:33:53.731150 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-p744p" event={"ID":"f5acada6-99e9-4109-b6fe-0b636e85a756","Type":"ContainerStarted","Data":"97227c3e16bdc36f994f5f805ddf9fd10fb69469b9539838c257329919d2317a"} Dec 04 10:33:56 crc kubenswrapper[4943]: I1204 10:33:56.755277 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-p744p" event={"ID":"f5acada6-99e9-4109-b6fe-0b636e85a756","Type":"ContainerStarted","Data":"3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6"} Dec 04 10:33:56 crc kubenswrapper[4943]: I1204 10:33:56.777289 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-p744p" podStartSLOduration=3.909307241 podStartE2EDuration="4.777268817s" podCreationTimestamp="2025-12-04 10:33:52 +0000 UTC" firstStartedPulling="2025-12-04 10:33:53.315495523 +0000 UTC m=+1081.904471391" lastFinishedPulling="2025-12-04 10:33:54.183457099 +0000 UTC m=+1082.772432967" observedRunningTime="2025-12-04 10:33:56.771924842 +0000 UTC m=+1085.360900730" watchObservedRunningTime="2025-12-04 10:33:56.777268817 +0000 UTC m=+1085.366244685" Dec 04 10:34:01 crc kubenswrapper[4943]: I1204 10:34:01.804721 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"057bf926-e2de-427f-9288-345fa9a798a6","Type":"ContainerStarted","Data":"04af2f59ff3e95f895fbde1e50874174d5ef7263e4b181a70b8a4da6607535f2"} Dec 04 10:34:03 crc kubenswrapper[4943]: I1204 10:34:03.064457 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-p744p" Dec 04 10:34:03 crc kubenswrapper[4943]: I1204 10:34:03.064629 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-p744p" Dec 04 10:34:03 crc kubenswrapper[4943]: I1204 10:34:03.089777 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-p744p" Dec 04 10:34:03 crc kubenswrapper[4943]: I1204 10:34:03.837404 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-p744p" Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.407235 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp"] Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.409010 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.410921 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6gzwg" Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.418754 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp"] Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.574177 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2391487b-52c7-4e1f-a03b-b20371f30445-bundle\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp\" (UID: \"2391487b-52c7-4e1f-a03b-b20371f30445\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.574249 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsbjj\" (UniqueName: \"kubernetes.io/projected/2391487b-52c7-4e1f-a03b-b20371f30445-kube-api-access-tsbjj\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp\" (UID: \"2391487b-52c7-4e1f-a03b-b20371f30445\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.574362 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2391487b-52c7-4e1f-a03b-b20371f30445-util\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp\" (UID: \"2391487b-52c7-4e1f-a03b-b20371f30445\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.675323 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2391487b-52c7-4e1f-a03b-b20371f30445-bundle\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp\" (UID: \"2391487b-52c7-4e1f-a03b-b20371f30445\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.675378 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsbjj\" (UniqueName: \"kubernetes.io/projected/2391487b-52c7-4e1f-a03b-b20371f30445-kube-api-access-tsbjj\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp\" (UID: \"2391487b-52c7-4e1f-a03b-b20371f30445\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.675456 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2391487b-52c7-4e1f-a03b-b20371f30445-util\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp\" (UID: \"2391487b-52c7-4e1f-a03b-b20371f30445\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.675945 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2391487b-52c7-4e1f-a03b-b20371f30445-bundle\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp\" (UID: \"2391487b-52c7-4e1f-a03b-b20371f30445\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.675957 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2391487b-52c7-4e1f-a03b-b20371f30445-util\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp\" (UID: \"2391487b-52c7-4e1f-a03b-b20371f30445\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.694371 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsbjj\" (UniqueName: \"kubernetes.io/projected/2391487b-52c7-4e1f-a03b-b20371f30445-kube-api-access-tsbjj\") pod \"49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp\" (UID: \"2391487b-52c7-4e1f-a03b-b20371f30445\") " pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:05 crc kubenswrapper[4943]: I1204 10:34:05.726791 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:06 crc kubenswrapper[4943]: I1204 10:34:06.149588 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp"] Dec 04 10:34:06 crc kubenswrapper[4943]: I1204 10:34:06.838284 4943 generic.go:334] "Generic (PLEG): container finished" podID="2391487b-52c7-4e1f-a03b-b20371f30445" containerID="dc54ccd7fcdb9d90906924dc9830ffe2f2a3c3ba6c9237079ceb16bf5006727c" exitCode=0 Dec 04 10:34:06 crc kubenswrapper[4943]: I1204 10:34:06.838405 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" event={"ID":"2391487b-52c7-4e1f-a03b-b20371f30445","Type":"ContainerDied","Data":"dc54ccd7fcdb9d90906924dc9830ffe2f2a3c3ba6c9237079ceb16bf5006727c"} Dec 04 10:34:06 crc kubenswrapper[4943]: I1204 10:34:06.838657 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" event={"ID":"2391487b-52c7-4e1f-a03b-b20371f30445","Type":"ContainerStarted","Data":"0e93f8d89a8bf9b2cb7121401333160b1ab1f1c0d880b8a5ac34ecf8d5f64a22"} Dec 04 10:34:07 crc kubenswrapper[4943]: I1204 10:34:07.853261 4943 generic.go:334] "Generic (PLEG): container finished" podID="2391487b-52c7-4e1f-a03b-b20371f30445" containerID="eef8d03406a95dc991fb6f7c8dc1584719b2b49f9ded87299333b8d6815c24ac" exitCode=0 Dec 04 10:34:07 crc kubenswrapper[4943]: I1204 10:34:07.853665 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" event={"ID":"2391487b-52c7-4e1f-a03b-b20371f30445","Type":"ContainerDied","Data":"eef8d03406a95dc991fb6f7c8dc1584719b2b49f9ded87299333b8d6815c24ac"} Dec 04 10:34:08 crc kubenswrapper[4943]: I1204 10:34:08.863694 4943 generic.go:334] "Generic (PLEG): container finished" podID="2391487b-52c7-4e1f-a03b-b20371f30445" containerID="d2390b136890132f4569ff5ff2cb50b6fad4045793a4097fc9cef7eba14c4667" exitCode=0 Dec 04 10:34:08 crc kubenswrapper[4943]: I1204 10:34:08.863742 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" event={"ID":"2391487b-52c7-4e1f-a03b-b20371f30445","Type":"ContainerDied","Data":"d2390b136890132f4569ff5ff2cb50b6fad4045793a4097fc9cef7eba14c4667"} Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.130758 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.245630 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsbjj\" (UniqueName: \"kubernetes.io/projected/2391487b-52c7-4e1f-a03b-b20371f30445-kube-api-access-tsbjj\") pod \"2391487b-52c7-4e1f-a03b-b20371f30445\" (UID: \"2391487b-52c7-4e1f-a03b-b20371f30445\") " Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.245707 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2391487b-52c7-4e1f-a03b-b20371f30445-util\") pod \"2391487b-52c7-4e1f-a03b-b20371f30445\" (UID: \"2391487b-52c7-4e1f-a03b-b20371f30445\") " Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.245831 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2391487b-52c7-4e1f-a03b-b20371f30445-bundle\") pod \"2391487b-52c7-4e1f-a03b-b20371f30445\" (UID: \"2391487b-52c7-4e1f-a03b-b20371f30445\") " Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.246731 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2391487b-52c7-4e1f-a03b-b20371f30445-bundle" (OuterVolumeSpecName: "bundle") pod "2391487b-52c7-4e1f-a03b-b20371f30445" (UID: "2391487b-52c7-4e1f-a03b-b20371f30445"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.251541 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2391487b-52c7-4e1f-a03b-b20371f30445-kube-api-access-tsbjj" (OuterVolumeSpecName: "kube-api-access-tsbjj") pod "2391487b-52c7-4e1f-a03b-b20371f30445" (UID: "2391487b-52c7-4e1f-a03b-b20371f30445"). InnerVolumeSpecName "kube-api-access-tsbjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.260344 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2391487b-52c7-4e1f-a03b-b20371f30445-util" (OuterVolumeSpecName: "util") pod "2391487b-52c7-4e1f-a03b-b20371f30445" (UID: "2391487b-52c7-4e1f-a03b-b20371f30445"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.347153 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2391487b-52c7-4e1f-a03b-b20371f30445-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.347213 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsbjj\" (UniqueName: \"kubernetes.io/projected/2391487b-52c7-4e1f-a03b-b20371f30445-kube-api-access-tsbjj\") on node \"crc\" DevicePath \"\"" Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.347233 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2391487b-52c7-4e1f-a03b-b20371f30445-util\") on node \"crc\" DevicePath \"\"" Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.877859 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" event={"ID":"2391487b-52c7-4e1f-a03b-b20371f30445","Type":"ContainerDied","Data":"0e93f8d89a8bf9b2cb7121401333160b1ab1f1c0d880b8a5ac34ecf8d5f64a22"} Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.877929 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e93f8d89a8bf9b2cb7121401333160b1ab1f1c0d880b8a5ac34ecf8d5f64a22" Dec 04 10:34:10 crc kubenswrapper[4943]: I1204 10:34:10.878026 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.386329 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7"] Dec 04 10:34:17 crc kubenswrapper[4943]: E1204 10:34:17.387051 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2391487b-52c7-4e1f-a03b-b20371f30445" containerName="pull" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.387076 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="2391487b-52c7-4e1f-a03b-b20371f30445" containerName="pull" Dec 04 10:34:17 crc kubenswrapper[4943]: E1204 10:34:17.387094 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2391487b-52c7-4e1f-a03b-b20371f30445" containerName="util" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.387100 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="2391487b-52c7-4e1f-a03b-b20371f30445" containerName="util" Dec 04 10:34:17 crc kubenswrapper[4943]: E1204 10:34:17.387116 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2391487b-52c7-4e1f-a03b-b20371f30445" containerName="extract" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.387124 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="2391487b-52c7-4e1f-a03b-b20371f30445" containerName="extract" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.387264 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="2391487b-52c7-4e1f-a03b-b20371f30445" containerName="extract" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.387686 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.389982 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.391621 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-2gszl" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.398506 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6006a85-81c2-4909-912e-16c49c862f7a-webhook-cert\") pod \"keystone-operator-controller-manager-794886f574-lzdr7\" (UID: \"f6006a85-81c2-4909-912e-16c49c862f7a\") " pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.398573 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6drc\" (UniqueName: \"kubernetes.io/projected/f6006a85-81c2-4909-912e-16c49c862f7a-kube-api-access-z6drc\") pod \"keystone-operator-controller-manager-794886f574-lzdr7\" (UID: \"f6006a85-81c2-4909-912e-16c49c862f7a\") " pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.398604 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6006a85-81c2-4909-912e-16c49c862f7a-apiservice-cert\") pod \"keystone-operator-controller-manager-794886f574-lzdr7\" (UID: \"f6006a85-81c2-4909-912e-16c49c862f7a\") " pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.403064 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7"] Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.499687 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6006a85-81c2-4909-912e-16c49c862f7a-webhook-cert\") pod \"keystone-operator-controller-manager-794886f574-lzdr7\" (UID: \"f6006a85-81c2-4909-912e-16c49c862f7a\") " pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.499914 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6drc\" (UniqueName: \"kubernetes.io/projected/f6006a85-81c2-4909-912e-16c49c862f7a-kube-api-access-z6drc\") pod \"keystone-operator-controller-manager-794886f574-lzdr7\" (UID: \"f6006a85-81c2-4909-912e-16c49c862f7a\") " pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.499950 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6006a85-81c2-4909-912e-16c49c862f7a-apiservice-cert\") pod \"keystone-operator-controller-manager-794886f574-lzdr7\" (UID: \"f6006a85-81c2-4909-912e-16c49c862f7a\") " pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.505179 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6006a85-81c2-4909-912e-16c49c862f7a-webhook-cert\") pod \"keystone-operator-controller-manager-794886f574-lzdr7\" (UID: \"f6006a85-81c2-4909-912e-16c49c862f7a\") " pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.505190 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6006a85-81c2-4909-912e-16c49c862f7a-apiservice-cert\") pod \"keystone-operator-controller-manager-794886f574-lzdr7\" (UID: \"f6006a85-81c2-4909-912e-16c49c862f7a\") " pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.522020 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6drc\" (UniqueName: \"kubernetes.io/projected/f6006a85-81c2-4909-912e-16c49c862f7a-kube-api-access-z6drc\") pod \"keystone-operator-controller-manager-794886f574-lzdr7\" (UID: \"f6006a85-81c2-4909-912e-16c49c862f7a\") " pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.707219 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:17 crc kubenswrapper[4943]: I1204 10:34:17.958456 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7"] Dec 04 10:34:18 crc kubenswrapper[4943]: I1204 10:34:18.954469 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" event={"ID":"f6006a85-81c2-4909-912e-16c49c862f7a","Type":"ContainerStarted","Data":"296c2f2b7582287bb4102cf997c45adb71085a3e41dcebf80012c6b2005aa4fc"} Dec 04 10:34:23 crc kubenswrapper[4943]: I1204 10:34:23.007620 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" event={"ID":"f6006a85-81c2-4909-912e-16c49c862f7a","Type":"ContainerStarted","Data":"7bd46b939209cfe6a74549112d929c167c3310fc4fc3f47953a37260e2b66125"} Dec 04 10:34:23 crc kubenswrapper[4943]: I1204 10:34:23.008220 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:27 crc kubenswrapper[4943]: I1204 10:34:27.715419 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:34:27 crc kubenswrapper[4943]: I1204 10:34:27.777555 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" podStartSLOduration=6.01812498 podStartE2EDuration="10.777507045s" podCreationTimestamp="2025-12-04 10:34:17 +0000 UTC" firstStartedPulling="2025-12-04 10:34:17.965347979 +0000 UTC m=+1106.554323847" lastFinishedPulling="2025-12-04 10:34:22.724730044 +0000 UTC m=+1111.313705912" observedRunningTime="2025-12-04 10:34:23.02574245 +0000 UTC m=+1111.614718328" watchObservedRunningTime="2025-12-04 10:34:27.777507045 +0000 UTC m=+1116.366482923" Dec 04 10:34:33 crc kubenswrapper[4943]: I1204 10:34:33.069517 4943 generic.go:334] "Generic (PLEG): container finished" podID="057bf926-e2de-427f-9288-345fa9a798a6" containerID="04af2f59ff3e95f895fbde1e50874174d5ef7263e4b181a70b8a4da6607535f2" exitCode=0 Dec 04 10:34:33 crc kubenswrapper[4943]: I1204 10:34:33.070300 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"057bf926-e2de-427f-9288-345fa9a798a6","Type":"ContainerDied","Data":"04af2f59ff3e95f895fbde1e50874174d5ef7263e4b181a70b8a4da6607535f2"} Dec 04 10:34:34 crc kubenswrapper[4943]: I1204 10:34:34.078987 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"057bf926-e2de-427f-9288-345fa9a798a6","Type":"ContainerStarted","Data":"1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7"} Dec 04 10:34:34 crc kubenswrapper[4943]: I1204 10:34:34.079511 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:34:39 crc kubenswrapper[4943]: I1204 10:34:39.333455 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/rabbitmq-server-0" podStartSLOduration=42.638319596 podStartE2EDuration="49.333438057s" podCreationTimestamp="2025-12-04 10:33:50 +0000 UTC" firstStartedPulling="2025-12-04 10:33:51.857047153 +0000 UTC m=+1080.446023021" lastFinishedPulling="2025-12-04 10:33:58.552165614 +0000 UTC m=+1087.141141482" observedRunningTime="2025-12-04 10:34:34.101799034 +0000 UTC m=+1122.690774912" watchObservedRunningTime="2025-12-04 10:34:39.333438057 +0000 UTC m=+1127.922413925" Dec 04 10:34:39 crc kubenswrapper[4943]: I1204 10:34:39.338074 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-index-qdncn"] Dec 04 10:34:39 crc kubenswrapper[4943]: I1204 10:34:39.339042 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-qdncn" Dec 04 10:34:39 crc kubenswrapper[4943]: I1204 10:34:39.342842 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-index-dockercfg-r9vbq" Dec 04 10:34:39 crc kubenswrapper[4943]: I1204 10:34:39.363535 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-index-qdncn"] Dec 04 10:34:39 crc kubenswrapper[4943]: I1204 10:34:39.429911 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpz7t\" (UniqueName: \"kubernetes.io/projected/ccd80df8-e6b5-4714-a462-fce5c2b6af85-kube-api-access-gpz7t\") pod \"cinder-operator-index-qdncn\" (UID: \"ccd80df8-e6b5-4714-a462-fce5c2b6af85\") " pod="openstack-operators/cinder-operator-index-qdncn" Dec 04 10:34:39 crc kubenswrapper[4943]: I1204 10:34:39.530795 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpz7t\" (UniqueName: \"kubernetes.io/projected/ccd80df8-e6b5-4714-a462-fce5c2b6af85-kube-api-access-gpz7t\") pod \"cinder-operator-index-qdncn\" (UID: \"ccd80df8-e6b5-4714-a462-fce5c2b6af85\") " pod="openstack-operators/cinder-operator-index-qdncn" Dec 04 10:34:39 crc kubenswrapper[4943]: I1204 10:34:39.560537 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpz7t\" (UniqueName: \"kubernetes.io/projected/ccd80df8-e6b5-4714-a462-fce5c2b6af85-kube-api-access-gpz7t\") pod \"cinder-operator-index-qdncn\" (UID: \"ccd80df8-e6b5-4714-a462-fce5c2b6af85\") " pod="openstack-operators/cinder-operator-index-qdncn" Dec 04 10:34:39 crc kubenswrapper[4943]: I1204 10:34:39.663911 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-qdncn" Dec 04 10:34:40 crc kubenswrapper[4943]: I1204 10:34:40.061615 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-index-qdncn"] Dec 04 10:34:40 crc kubenswrapper[4943]: I1204 10:34:40.114662 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-qdncn" event={"ID":"ccd80df8-e6b5-4714-a462-fce5c2b6af85","Type":"ContainerStarted","Data":"559a16573294063b207a10d73ec01316e7a7d688aacbec9e5d06ace47e8b4aa6"} Dec 04 10:34:42 crc kubenswrapper[4943]: I1204 10:34:42.127771 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-qdncn" event={"ID":"ccd80df8-e6b5-4714-a462-fce5c2b6af85","Type":"ContainerStarted","Data":"07400d6f43ac162265f1754d269322529733bda95ac6ecd55d24c4bcdda2e175"} Dec 04 10:34:42 crc kubenswrapper[4943]: I1204 10:34:42.146569 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-index-qdncn" podStartSLOduration=1.88057666 podStartE2EDuration="3.146542965s" podCreationTimestamp="2025-12-04 10:34:39 +0000 UTC" firstStartedPulling="2025-12-04 10:34:40.065127514 +0000 UTC m=+1128.654103382" lastFinishedPulling="2025-12-04 10:34:41.331093819 +0000 UTC m=+1129.920069687" observedRunningTime="2025-12-04 10:34:42.140766126 +0000 UTC m=+1130.729741994" watchObservedRunningTime="2025-12-04 10:34:42.146542965 +0000 UTC m=+1130.735518833" Dec 04 10:34:48 crc kubenswrapper[4943]: I1204 10:34:48.630320 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:34:48 crc kubenswrapper[4943]: I1204 10:34:48.630955 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:34:49 crc kubenswrapper[4943]: I1204 10:34:49.664265 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-index-qdncn" Dec 04 10:34:49 crc kubenswrapper[4943]: I1204 10:34:49.664563 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/cinder-operator-index-qdncn" Dec 04 10:34:49 crc kubenswrapper[4943]: I1204 10:34:49.686359 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/cinder-operator-index-qdncn" Dec 04 10:34:50 crc kubenswrapper[4943]: I1204 10:34:50.205510 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-index-qdncn" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.083371 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j"] Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.084854 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.090853 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-db-create-dndb4"] Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.091955 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-dndb4" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.092847 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-db-secret" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.100315 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j"] Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.105749 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-dndb4"] Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.248286 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/954e673d-85ec-4ebd-9c7a-105986c8ccab-operator-scripts\") pod \"keystone-fea1-account-create-update-gpj8j\" (UID: \"954e673d-85ec-4ebd-9c7a-105986c8ccab\") " pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.248816 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96b5j\" (UniqueName: \"kubernetes.io/projected/954e673d-85ec-4ebd-9c7a-105986c8ccab-kube-api-access-96b5j\") pod \"keystone-fea1-account-create-update-gpj8j\" (UID: \"954e673d-85ec-4ebd-9c7a-105986c8ccab\") " pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.249041 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vptp2\" (UniqueName: \"kubernetes.io/projected/a6266a3a-7624-4cd4-b9de-fb806517a35c-kube-api-access-vptp2\") pod \"keystone-db-create-dndb4\" (UID: \"a6266a3a-7624-4cd4-b9de-fb806517a35c\") " pod="cinder-kuttl-tests/keystone-db-create-dndb4" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.249102 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6266a3a-7624-4cd4-b9de-fb806517a35c-operator-scripts\") pod \"keystone-db-create-dndb4\" (UID: \"a6266a3a-7624-4cd4-b9de-fb806517a35c\") " pod="cinder-kuttl-tests/keystone-db-create-dndb4" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.350283 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/954e673d-85ec-4ebd-9c7a-105986c8ccab-operator-scripts\") pod \"keystone-fea1-account-create-update-gpj8j\" (UID: \"954e673d-85ec-4ebd-9c7a-105986c8ccab\") " pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.350698 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96b5j\" (UniqueName: \"kubernetes.io/projected/954e673d-85ec-4ebd-9c7a-105986c8ccab-kube-api-access-96b5j\") pod \"keystone-fea1-account-create-update-gpj8j\" (UID: \"954e673d-85ec-4ebd-9c7a-105986c8ccab\") " pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.350971 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vptp2\" (UniqueName: \"kubernetes.io/projected/a6266a3a-7624-4cd4-b9de-fb806517a35c-kube-api-access-vptp2\") pod \"keystone-db-create-dndb4\" (UID: \"a6266a3a-7624-4cd4-b9de-fb806517a35c\") " pod="cinder-kuttl-tests/keystone-db-create-dndb4" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.351279 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/954e673d-85ec-4ebd-9c7a-105986c8ccab-operator-scripts\") pod \"keystone-fea1-account-create-update-gpj8j\" (UID: \"954e673d-85ec-4ebd-9c7a-105986c8ccab\") " pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.351413 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6266a3a-7624-4cd4-b9de-fb806517a35c-operator-scripts\") pod \"keystone-db-create-dndb4\" (UID: \"a6266a3a-7624-4cd4-b9de-fb806517a35c\") " pod="cinder-kuttl-tests/keystone-db-create-dndb4" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.352675 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6266a3a-7624-4cd4-b9de-fb806517a35c-operator-scripts\") pod \"keystone-db-create-dndb4\" (UID: \"a6266a3a-7624-4cd4-b9de-fb806517a35c\") " pod="cinder-kuttl-tests/keystone-db-create-dndb4" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.374923 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vptp2\" (UniqueName: \"kubernetes.io/projected/a6266a3a-7624-4cd4-b9de-fb806517a35c-kube-api-access-vptp2\") pod \"keystone-db-create-dndb4\" (UID: \"a6266a3a-7624-4cd4-b9de-fb806517a35c\") " pod="cinder-kuttl-tests/keystone-db-create-dndb4" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.375860 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96b5j\" (UniqueName: \"kubernetes.io/projected/954e673d-85ec-4ebd-9c7a-105986c8ccab-kube-api-access-96b5j\") pod \"keystone-fea1-account-create-update-gpj8j\" (UID: \"954e673d-85ec-4ebd-9c7a-105986c8ccab\") " pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.417433 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.422626 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.435003 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-dndb4" Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.718482 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-dndb4"] Dec 04 10:34:51 crc kubenswrapper[4943]: I1204 10:34:51.885179 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j"] Dec 04 10:34:51 crc kubenswrapper[4943]: W1204 10:34:51.888138 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod954e673d_85ec_4ebd_9c7a_105986c8ccab.slice/crio-f1192872f6b4f96299cc5f1637cbdcbae46ecfeaa2904e262ac1e9c6f16d7e97 WatchSource:0}: Error finding container f1192872f6b4f96299cc5f1637cbdcbae46ecfeaa2904e262ac1e9c6f16d7e97: Status 404 returned error can't find the container with id f1192872f6b4f96299cc5f1637cbdcbae46ecfeaa2904e262ac1e9c6f16d7e97 Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.177810 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv"] Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.179166 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.181007 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6gzwg" Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.192745 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv"] Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.193815 4943 generic.go:334] "Generic (PLEG): container finished" podID="a6266a3a-7624-4cd4-b9de-fb806517a35c" containerID="1e42cff7bee1e980d96bb2da3bc885ca5a59eefff74bc4aee1b1033dfea51910" exitCode=0 Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.193873 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-dndb4" event={"ID":"a6266a3a-7624-4cd4-b9de-fb806517a35c","Type":"ContainerDied","Data":"1e42cff7bee1e980d96bb2da3bc885ca5a59eefff74bc4aee1b1033dfea51910"} Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.193900 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-dndb4" event={"ID":"a6266a3a-7624-4cd4-b9de-fb806517a35c","Type":"ContainerStarted","Data":"d464b6d1ab94e85561556a2bf46d47a63850d45c1849dcdd558630e199365cc4"} Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.196071 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" event={"ID":"954e673d-85ec-4ebd-9c7a-105986c8ccab","Type":"ContainerStarted","Data":"81d9788be38c3eac800e4924c843ed26d147a2ad0600cfa557d270575a37fa25"} Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.196117 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" event={"ID":"954e673d-85ec-4ebd-9c7a-105986c8ccab","Type":"ContainerStarted","Data":"f1192872f6b4f96299cc5f1637cbdcbae46ecfeaa2904e262ac1e9c6f16d7e97"} Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.219546 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" podStartSLOduration=1.219523682 podStartE2EDuration="1.219523682s" podCreationTimestamp="2025-12-04 10:34:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:34:52.218224546 +0000 UTC m=+1140.807200414" watchObservedRunningTime="2025-12-04 10:34:52.219523682 +0000 UTC m=+1140.808499550" Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.373870 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3349476a-f5bc-40ce-a663-561662c8a6ac-util\") pod \"ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv\" (UID: \"3349476a-f5bc-40ce-a663-561662c8a6ac\") " pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.373970 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3349476a-f5bc-40ce-a663-561662c8a6ac-bundle\") pod \"ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv\" (UID: \"3349476a-f5bc-40ce-a663-561662c8a6ac\") " pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.374014 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hj7z\" (UniqueName: \"kubernetes.io/projected/3349476a-f5bc-40ce-a663-561662c8a6ac-kube-api-access-4hj7z\") pod \"ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv\" (UID: \"3349476a-f5bc-40ce-a663-561662c8a6ac\") " pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.474873 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hj7z\" (UniqueName: \"kubernetes.io/projected/3349476a-f5bc-40ce-a663-561662c8a6ac-kube-api-access-4hj7z\") pod \"ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv\" (UID: \"3349476a-f5bc-40ce-a663-561662c8a6ac\") " pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.474998 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3349476a-f5bc-40ce-a663-561662c8a6ac-util\") pod \"ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv\" (UID: \"3349476a-f5bc-40ce-a663-561662c8a6ac\") " pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.475056 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3349476a-f5bc-40ce-a663-561662c8a6ac-bundle\") pod \"ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv\" (UID: \"3349476a-f5bc-40ce-a663-561662c8a6ac\") " pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.475429 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3349476a-f5bc-40ce-a663-561662c8a6ac-util\") pod \"ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv\" (UID: \"3349476a-f5bc-40ce-a663-561662c8a6ac\") " pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.475545 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3349476a-f5bc-40ce-a663-561662c8a6ac-bundle\") pod \"ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv\" (UID: \"3349476a-f5bc-40ce-a663-561662c8a6ac\") " pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.500127 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hj7z\" (UniqueName: \"kubernetes.io/projected/3349476a-f5bc-40ce-a663-561662c8a6ac-kube-api-access-4hj7z\") pod \"ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv\" (UID: \"3349476a-f5bc-40ce-a663-561662c8a6ac\") " pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:52 crc kubenswrapper[4943]: I1204 10:34:52.791915 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:53 crc kubenswrapper[4943]: I1204 10:34:53.184490 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv"] Dec 04 10:34:53 crc kubenswrapper[4943]: W1204 10:34:53.189376 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3349476a_f5bc_40ce_a663_561662c8a6ac.slice/crio-f49dc5401a6edb7a6140d348c7017ed6f00dbcace1b5e99102b834060deccd40 WatchSource:0}: Error finding container f49dc5401a6edb7a6140d348c7017ed6f00dbcace1b5e99102b834060deccd40: Status 404 returned error can't find the container with id f49dc5401a6edb7a6140d348c7017ed6f00dbcace1b5e99102b834060deccd40 Dec 04 10:34:53 crc kubenswrapper[4943]: I1204 10:34:53.203968 4943 generic.go:334] "Generic (PLEG): container finished" podID="954e673d-85ec-4ebd-9c7a-105986c8ccab" containerID="81d9788be38c3eac800e4924c843ed26d147a2ad0600cfa557d270575a37fa25" exitCode=0 Dec 04 10:34:53 crc kubenswrapper[4943]: I1204 10:34:53.204167 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" event={"ID":"954e673d-85ec-4ebd-9c7a-105986c8ccab","Type":"ContainerDied","Data":"81d9788be38c3eac800e4924c843ed26d147a2ad0600cfa557d270575a37fa25"} Dec 04 10:34:53 crc kubenswrapper[4943]: I1204 10:34:53.205609 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" event={"ID":"3349476a-f5bc-40ce-a663-561662c8a6ac","Type":"ContainerStarted","Data":"f49dc5401a6edb7a6140d348c7017ed6f00dbcace1b5e99102b834060deccd40"} Dec 04 10:34:53 crc kubenswrapper[4943]: I1204 10:34:53.550000 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-dndb4" Dec 04 10:34:53 crc kubenswrapper[4943]: I1204 10:34:53.718213 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vptp2\" (UniqueName: \"kubernetes.io/projected/a6266a3a-7624-4cd4-b9de-fb806517a35c-kube-api-access-vptp2\") pod \"a6266a3a-7624-4cd4-b9de-fb806517a35c\" (UID: \"a6266a3a-7624-4cd4-b9de-fb806517a35c\") " Dec 04 10:34:53 crc kubenswrapper[4943]: I1204 10:34:53.718336 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6266a3a-7624-4cd4-b9de-fb806517a35c-operator-scripts\") pod \"a6266a3a-7624-4cd4-b9de-fb806517a35c\" (UID: \"a6266a3a-7624-4cd4-b9de-fb806517a35c\") " Dec 04 10:34:53 crc kubenswrapper[4943]: I1204 10:34:53.719029 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6266a3a-7624-4cd4-b9de-fb806517a35c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a6266a3a-7624-4cd4-b9de-fb806517a35c" (UID: "a6266a3a-7624-4cd4-b9de-fb806517a35c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:34:53 crc kubenswrapper[4943]: I1204 10:34:53.726363 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6266a3a-7624-4cd4-b9de-fb806517a35c-kube-api-access-vptp2" (OuterVolumeSpecName: "kube-api-access-vptp2") pod "a6266a3a-7624-4cd4-b9de-fb806517a35c" (UID: "a6266a3a-7624-4cd4-b9de-fb806517a35c"). InnerVolumeSpecName "kube-api-access-vptp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:34:53 crc kubenswrapper[4943]: I1204 10:34:53.820751 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vptp2\" (UniqueName: \"kubernetes.io/projected/a6266a3a-7624-4cd4-b9de-fb806517a35c-kube-api-access-vptp2\") on node \"crc\" DevicePath \"\"" Dec 04 10:34:53 crc kubenswrapper[4943]: I1204 10:34:53.820788 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6266a3a-7624-4cd4-b9de-fb806517a35c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:34:54 crc kubenswrapper[4943]: I1204 10:34:54.213348 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-dndb4" Dec 04 10:34:54 crc kubenswrapper[4943]: I1204 10:34:54.213347 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-dndb4" event={"ID":"a6266a3a-7624-4cd4-b9de-fb806517a35c","Type":"ContainerDied","Data":"d464b6d1ab94e85561556a2bf46d47a63850d45c1849dcdd558630e199365cc4"} Dec 04 10:34:54 crc kubenswrapper[4943]: I1204 10:34:54.213458 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d464b6d1ab94e85561556a2bf46d47a63850d45c1849dcdd558630e199365cc4" Dec 04 10:34:54 crc kubenswrapper[4943]: I1204 10:34:54.215164 4943 generic.go:334] "Generic (PLEG): container finished" podID="3349476a-f5bc-40ce-a663-561662c8a6ac" containerID="2bc153a1c3c30c517d39101699cd07667061e44a498b4d660f6d8f6bc96fee74" exitCode=0 Dec 04 10:34:54 crc kubenswrapper[4943]: I1204 10:34:54.215272 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" event={"ID":"3349476a-f5bc-40ce-a663-561662c8a6ac","Type":"ContainerDied","Data":"2bc153a1c3c30c517d39101699cd07667061e44a498b4d660f6d8f6bc96fee74"} Dec 04 10:34:54 crc kubenswrapper[4943]: I1204 10:34:54.771380 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" Dec 04 10:34:54 crc kubenswrapper[4943]: I1204 10:34:54.872453 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/954e673d-85ec-4ebd-9c7a-105986c8ccab-operator-scripts\") pod \"954e673d-85ec-4ebd-9c7a-105986c8ccab\" (UID: \"954e673d-85ec-4ebd-9c7a-105986c8ccab\") " Dec 04 10:34:54 crc kubenswrapper[4943]: I1204 10:34:54.872533 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96b5j\" (UniqueName: \"kubernetes.io/projected/954e673d-85ec-4ebd-9c7a-105986c8ccab-kube-api-access-96b5j\") pod \"954e673d-85ec-4ebd-9c7a-105986c8ccab\" (UID: \"954e673d-85ec-4ebd-9c7a-105986c8ccab\") " Dec 04 10:34:54 crc kubenswrapper[4943]: I1204 10:34:54.873903 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/954e673d-85ec-4ebd-9c7a-105986c8ccab-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "954e673d-85ec-4ebd-9c7a-105986c8ccab" (UID: "954e673d-85ec-4ebd-9c7a-105986c8ccab"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:34:54 crc kubenswrapper[4943]: I1204 10:34:54.877380 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/954e673d-85ec-4ebd-9c7a-105986c8ccab-kube-api-access-96b5j" (OuterVolumeSpecName: "kube-api-access-96b5j") pod "954e673d-85ec-4ebd-9c7a-105986c8ccab" (UID: "954e673d-85ec-4ebd-9c7a-105986c8ccab"). InnerVolumeSpecName "kube-api-access-96b5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:34:54 crc kubenswrapper[4943]: I1204 10:34:54.973996 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/954e673d-85ec-4ebd-9c7a-105986c8ccab-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:34:54 crc kubenswrapper[4943]: I1204 10:34:54.974032 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96b5j\" (UniqueName: \"kubernetes.io/projected/954e673d-85ec-4ebd-9c7a-105986c8ccab-kube-api-access-96b5j\") on node \"crc\" DevicePath \"\"" Dec 04 10:34:55 crc kubenswrapper[4943]: I1204 10:34:55.223777 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" Dec 04 10:34:55 crc kubenswrapper[4943]: I1204 10:34:55.223780 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j" event={"ID":"954e673d-85ec-4ebd-9c7a-105986c8ccab","Type":"ContainerDied","Data":"f1192872f6b4f96299cc5f1637cbdcbae46ecfeaa2904e262ac1e9c6f16d7e97"} Dec 04 10:34:55 crc kubenswrapper[4943]: I1204 10:34:55.223828 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1192872f6b4f96299cc5f1637cbdcbae46ecfeaa2904e262ac1e9c6f16d7e97" Dec 04 10:34:55 crc kubenswrapper[4943]: I1204 10:34:55.226033 4943 generic.go:334] "Generic (PLEG): container finished" podID="3349476a-f5bc-40ce-a663-561662c8a6ac" containerID="52589586ed94a0549a6570004e167d9ad64bca476673e1655a83ec0bcdefe3a4" exitCode=0 Dec 04 10:34:55 crc kubenswrapper[4943]: I1204 10:34:55.226082 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" event={"ID":"3349476a-f5bc-40ce-a663-561662c8a6ac","Type":"ContainerDied","Data":"52589586ed94a0549a6570004e167d9ad64bca476673e1655a83ec0bcdefe3a4"} Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.233826 4943 generic.go:334] "Generic (PLEG): container finished" podID="3349476a-f5bc-40ce-a663-561662c8a6ac" containerID="a69e6ccf43199cde628626de373ebac1c8bf085abb7dadd544d0c92c218a6b90" exitCode=0 Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.233865 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" event={"ID":"3349476a-f5bc-40ce-a663-561662c8a6ac","Type":"ContainerDied","Data":"a69e6ccf43199cde628626de373ebac1c8bf085abb7dadd544d0c92c218a6b90"} Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.634771 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-q94wl"] Dec 04 10:34:56 crc kubenswrapper[4943]: E1204 10:34:56.635078 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="954e673d-85ec-4ebd-9c7a-105986c8ccab" containerName="mariadb-account-create-update" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.635092 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="954e673d-85ec-4ebd-9c7a-105986c8ccab" containerName="mariadb-account-create-update" Dec 04 10:34:56 crc kubenswrapper[4943]: E1204 10:34:56.635107 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6266a3a-7624-4cd4-b9de-fb806517a35c" containerName="mariadb-database-create" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.635113 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6266a3a-7624-4cd4-b9de-fb806517a35c" containerName="mariadb-database-create" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.635325 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="954e673d-85ec-4ebd-9c7a-105986c8ccab" containerName="mariadb-account-create-update" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.635342 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6266a3a-7624-4cd4-b9de-fb806517a35c" containerName="mariadb-database-create" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.635795 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-q94wl" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.637653 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.637862 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-lc4gq" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.638177 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.638657 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.642712 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-q94wl"] Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.700026 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06955606-470a-4395-a345-42d5170fa271-config-data\") pod \"keystone-db-sync-q94wl\" (UID: \"06955606-470a-4395-a345-42d5170fa271\") " pod="cinder-kuttl-tests/keystone-db-sync-q94wl" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.700112 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lb8v\" (UniqueName: \"kubernetes.io/projected/06955606-470a-4395-a345-42d5170fa271-kube-api-access-8lb8v\") pod \"keystone-db-sync-q94wl\" (UID: \"06955606-470a-4395-a345-42d5170fa271\") " pod="cinder-kuttl-tests/keystone-db-sync-q94wl" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.801942 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06955606-470a-4395-a345-42d5170fa271-config-data\") pod \"keystone-db-sync-q94wl\" (UID: \"06955606-470a-4395-a345-42d5170fa271\") " pod="cinder-kuttl-tests/keystone-db-sync-q94wl" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.802083 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lb8v\" (UniqueName: \"kubernetes.io/projected/06955606-470a-4395-a345-42d5170fa271-kube-api-access-8lb8v\") pod \"keystone-db-sync-q94wl\" (UID: \"06955606-470a-4395-a345-42d5170fa271\") " pod="cinder-kuttl-tests/keystone-db-sync-q94wl" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.808528 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06955606-470a-4395-a345-42d5170fa271-config-data\") pod \"keystone-db-sync-q94wl\" (UID: \"06955606-470a-4395-a345-42d5170fa271\") " pod="cinder-kuttl-tests/keystone-db-sync-q94wl" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.823947 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lb8v\" (UniqueName: \"kubernetes.io/projected/06955606-470a-4395-a345-42d5170fa271-kube-api-access-8lb8v\") pod \"keystone-db-sync-q94wl\" (UID: \"06955606-470a-4395-a345-42d5170fa271\") " pod="cinder-kuttl-tests/keystone-db-sync-q94wl" Dec 04 10:34:56 crc kubenswrapper[4943]: I1204 10:34:56.949210 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-q94wl" Dec 04 10:34:57 crc kubenswrapper[4943]: I1204 10:34:57.341639 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-q94wl"] Dec 04 10:34:57 crc kubenswrapper[4943]: I1204 10:34:57.361402 4943 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 10:34:57 crc kubenswrapper[4943]: I1204 10:34:57.460062 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:57 crc kubenswrapper[4943]: I1204 10:34:57.614246 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3349476a-f5bc-40ce-a663-561662c8a6ac-bundle\") pod \"3349476a-f5bc-40ce-a663-561662c8a6ac\" (UID: \"3349476a-f5bc-40ce-a663-561662c8a6ac\") " Dec 04 10:34:57 crc kubenswrapper[4943]: I1204 10:34:57.614418 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hj7z\" (UniqueName: \"kubernetes.io/projected/3349476a-f5bc-40ce-a663-561662c8a6ac-kube-api-access-4hj7z\") pod \"3349476a-f5bc-40ce-a663-561662c8a6ac\" (UID: \"3349476a-f5bc-40ce-a663-561662c8a6ac\") " Dec 04 10:34:57 crc kubenswrapper[4943]: I1204 10:34:57.614493 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3349476a-f5bc-40ce-a663-561662c8a6ac-util\") pod \"3349476a-f5bc-40ce-a663-561662c8a6ac\" (UID: \"3349476a-f5bc-40ce-a663-561662c8a6ac\") " Dec 04 10:34:57 crc kubenswrapper[4943]: I1204 10:34:57.622067 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3349476a-f5bc-40ce-a663-561662c8a6ac-bundle" (OuterVolumeSpecName: "bundle") pod "3349476a-f5bc-40ce-a663-561662c8a6ac" (UID: "3349476a-f5bc-40ce-a663-561662c8a6ac"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:34:57 crc kubenswrapper[4943]: I1204 10:34:57.633554 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3349476a-f5bc-40ce-a663-561662c8a6ac-kube-api-access-4hj7z" (OuterVolumeSpecName: "kube-api-access-4hj7z") pod "3349476a-f5bc-40ce-a663-561662c8a6ac" (UID: "3349476a-f5bc-40ce-a663-561662c8a6ac"). InnerVolumeSpecName "kube-api-access-4hj7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:34:57 crc kubenswrapper[4943]: I1204 10:34:57.648163 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3349476a-f5bc-40ce-a663-561662c8a6ac-util" (OuterVolumeSpecName: "util") pod "3349476a-f5bc-40ce-a663-561662c8a6ac" (UID: "3349476a-f5bc-40ce-a663-561662c8a6ac"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:34:57 crc kubenswrapper[4943]: I1204 10:34:57.726971 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hj7z\" (UniqueName: \"kubernetes.io/projected/3349476a-f5bc-40ce-a663-561662c8a6ac-kube-api-access-4hj7z\") on node \"crc\" DevicePath \"\"" Dec 04 10:34:57 crc kubenswrapper[4943]: I1204 10:34:57.727009 4943 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3349476a-f5bc-40ce-a663-561662c8a6ac-util\") on node \"crc\" DevicePath \"\"" Dec 04 10:34:57 crc kubenswrapper[4943]: I1204 10:34:57.727019 4943 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3349476a-f5bc-40ce-a663-561662c8a6ac-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:34:58 crc kubenswrapper[4943]: I1204 10:34:58.253019 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" event={"ID":"3349476a-f5bc-40ce-a663-561662c8a6ac","Type":"ContainerDied","Data":"f49dc5401a6edb7a6140d348c7017ed6f00dbcace1b5e99102b834060deccd40"} Dec 04 10:34:58 crc kubenswrapper[4943]: I1204 10:34:58.253425 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f49dc5401a6edb7a6140d348c7017ed6f00dbcace1b5e99102b834060deccd40" Dec 04 10:34:58 crc kubenswrapper[4943]: I1204 10:34:58.253161 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv" Dec 04 10:34:58 crc kubenswrapper[4943]: I1204 10:34:58.254186 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-q94wl" event={"ID":"06955606-470a-4395-a345-42d5170fa271","Type":"ContainerStarted","Data":"1efc59d1285fe2e45e83fa2ebda5ba6229e21ea3cd4520af65c9c68dbca42a31"} Dec 04 10:35:05 crc kubenswrapper[4943]: I1204 10:35:05.299783 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-q94wl" event={"ID":"06955606-470a-4395-a345-42d5170fa271","Type":"ContainerStarted","Data":"272ee3e9542601fd57eaa015217a3befda4353aacea9cb30765e43e04c4d9ceb"} Dec 04 10:35:05 crc kubenswrapper[4943]: I1204 10:35:05.320040 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-db-sync-q94wl" podStartSLOduration=2.356903899 podStartE2EDuration="9.320015046s" podCreationTimestamp="2025-12-04 10:34:56 +0000 UTC" firstStartedPulling="2025-12-04 10:34:57.361126522 +0000 UTC m=+1145.950102390" lastFinishedPulling="2025-12-04 10:35:04.324237669 +0000 UTC m=+1152.913213537" observedRunningTime="2025-12-04 10:35:05.313279 +0000 UTC m=+1153.902254868" watchObservedRunningTime="2025-12-04 10:35:05.320015046 +0000 UTC m=+1153.908990914" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.079554 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28"] Dec 04 10:35:07 crc kubenswrapper[4943]: E1204 10:35:07.080515 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3349476a-f5bc-40ce-a663-561662c8a6ac" containerName="util" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.080534 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3349476a-f5bc-40ce-a663-561662c8a6ac" containerName="util" Dec 04 10:35:07 crc kubenswrapper[4943]: E1204 10:35:07.080560 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3349476a-f5bc-40ce-a663-561662c8a6ac" containerName="extract" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.080567 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3349476a-f5bc-40ce-a663-561662c8a6ac" containerName="extract" Dec 04 10:35:07 crc kubenswrapper[4943]: E1204 10:35:07.080579 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3349476a-f5bc-40ce-a663-561662c8a6ac" containerName="pull" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.080585 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3349476a-f5bc-40ce-a663-561662c8a6ac" containerName="pull" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.080734 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3349476a-f5bc-40ce-a663-561662c8a6ac" containerName="extract" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.081407 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.086570 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-service-cert" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.086965 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-b46dl" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.095387 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28"] Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.195404 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-apiservice-cert\") pod \"cinder-operator-controller-manager-5f5d978b7f-vlt28\" (UID: \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\") " pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.195475 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-webhook-cert\") pod \"cinder-operator-controller-manager-5f5d978b7f-vlt28\" (UID: \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\") " pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.195494 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4mp7\" (UniqueName: \"kubernetes.io/projected/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-kube-api-access-l4mp7\") pod \"cinder-operator-controller-manager-5f5d978b7f-vlt28\" (UID: \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\") " pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.297149 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-apiservice-cert\") pod \"cinder-operator-controller-manager-5f5d978b7f-vlt28\" (UID: \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\") " pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.297246 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-webhook-cert\") pod \"cinder-operator-controller-manager-5f5d978b7f-vlt28\" (UID: \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\") " pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.297267 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4mp7\" (UniqueName: \"kubernetes.io/projected/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-kube-api-access-l4mp7\") pod \"cinder-operator-controller-manager-5f5d978b7f-vlt28\" (UID: \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\") " pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.304881 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-apiservice-cert\") pod \"cinder-operator-controller-manager-5f5d978b7f-vlt28\" (UID: \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\") " pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.304894 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-webhook-cert\") pod \"cinder-operator-controller-manager-5f5d978b7f-vlt28\" (UID: \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\") " pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.318869 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4mp7\" (UniqueName: \"kubernetes.io/projected/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-kube-api-access-l4mp7\") pod \"cinder-operator-controller-manager-5f5d978b7f-vlt28\" (UID: \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\") " pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.406007 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:07 crc kubenswrapper[4943]: I1204 10:35:07.846749 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28"] Dec 04 10:35:08 crc kubenswrapper[4943]: I1204 10:35:08.323644 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" event={"ID":"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03","Type":"ContainerStarted","Data":"c141f7bcd8030f2a01ebc9c515394a6d35466299a120cea8518687a23cd57e49"} Dec 04 10:35:08 crc kubenswrapper[4943]: I1204 10:35:08.325360 4943 generic.go:334] "Generic (PLEG): container finished" podID="06955606-470a-4395-a345-42d5170fa271" containerID="272ee3e9542601fd57eaa015217a3befda4353aacea9cb30765e43e04c4d9ceb" exitCode=0 Dec 04 10:35:08 crc kubenswrapper[4943]: I1204 10:35:08.325402 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-q94wl" event={"ID":"06955606-470a-4395-a345-42d5170fa271","Type":"ContainerDied","Data":"272ee3e9542601fd57eaa015217a3befda4353aacea9cb30765e43e04c4d9ceb"} Dec 04 10:35:09 crc kubenswrapper[4943]: I1204 10:35:09.634523 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-q94wl" Dec 04 10:35:09 crc kubenswrapper[4943]: I1204 10:35:09.733626 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lb8v\" (UniqueName: \"kubernetes.io/projected/06955606-470a-4395-a345-42d5170fa271-kube-api-access-8lb8v\") pod \"06955606-470a-4395-a345-42d5170fa271\" (UID: \"06955606-470a-4395-a345-42d5170fa271\") " Dec 04 10:35:09 crc kubenswrapper[4943]: I1204 10:35:09.733691 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06955606-470a-4395-a345-42d5170fa271-config-data\") pod \"06955606-470a-4395-a345-42d5170fa271\" (UID: \"06955606-470a-4395-a345-42d5170fa271\") " Dec 04 10:35:09 crc kubenswrapper[4943]: I1204 10:35:09.739391 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06955606-470a-4395-a345-42d5170fa271-kube-api-access-8lb8v" (OuterVolumeSpecName: "kube-api-access-8lb8v") pod "06955606-470a-4395-a345-42d5170fa271" (UID: "06955606-470a-4395-a345-42d5170fa271"). InnerVolumeSpecName "kube-api-access-8lb8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:35:09 crc kubenswrapper[4943]: I1204 10:35:09.766459 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06955606-470a-4395-a345-42d5170fa271-config-data" (OuterVolumeSpecName: "config-data") pod "06955606-470a-4395-a345-42d5170fa271" (UID: "06955606-470a-4395-a345-42d5170fa271"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:35:09 crc kubenswrapper[4943]: I1204 10:35:09.835631 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lb8v\" (UniqueName: \"kubernetes.io/projected/06955606-470a-4395-a345-42d5170fa271-kube-api-access-8lb8v\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:09 crc kubenswrapper[4943]: I1204 10:35:09.835676 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06955606-470a-4395-a345-42d5170fa271-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.343289 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-q94wl" event={"ID":"06955606-470a-4395-a345-42d5170fa271","Type":"ContainerDied","Data":"1efc59d1285fe2e45e83fa2ebda5ba6229e21ea3cd4520af65c9c68dbca42a31"} Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.343330 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1efc59d1285fe2e45e83fa2ebda5ba6229e21ea3cd4520af65c9c68dbca42a31" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.343382 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-q94wl" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.348401 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" event={"ID":"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03","Type":"ContainerStarted","Data":"8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe"} Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.349337 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.368831 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" podStartSLOduration=2.005899082 podStartE2EDuration="3.368808468s" podCreationTimestamp="2025-12-04 10:35:07 +0000 UTC" firstStartedPulling="2025-12-04 10:35:07.856991584 +0000 UTC m=+1156.445967452" lastFinishedPulling="2025-12-04 10:35:09.21990096 +0000 UTC m=+1157.808876838" observedRunningTime="2025-12-04 10:35:10.364545601 +0000 UTC m=+1158.953521479" watchObservedRunningTime="2025-12-04 10:35:10.368808468 +0000 UTC m=+1158.957784336" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.531140 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-crh8b"] Dec 04 10:35:10 crc kubenswrapper[4943]: E1204 10:35:10.531435 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06955606-470a-4395-a345-42d5170fa271" containerName="keystone-db-sync" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.531448 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="06955606-470a-4395-a345-42d5170fa271" containerName="keystone-db-sync" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.531567 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="06955606-470a-4395-a345-42d5170fa271" containerName="keystone-db-sync" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.531974 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.534163 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.534415 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.534765 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Dec 04 10:35:10 crc kubenswrapper[4943]: W1204 10:35:10.534995 4943 reflector.go:561] object-"cinder-kuttl-tests"/"osp-secret": failed to list *v1.Secret: secrets "osp-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "cinder-kuttl-tests": no relationship found between node 'crc' and this object Dec 04 10:35:10 crc kubenswrapper[4943]: E1204 10:35:10.535066 4943 reflector.go:158] "Unhandled Error" err="object-\"cinder-kuttl-tests\"/\"osp-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"osp-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"cinder-kuttl-tests\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.535121 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-lc4gq" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.545088 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-scripts\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.545135 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxjjv\" (UniqueName: \"kubernetes.io/projected/eaf4529b-f939-41eb-b31e-372ae95e1d51-kube-api-access-lxjjv\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.545529 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-config-data\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.545648 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-fernet-keys\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.545680 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-credential-keys\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.545701 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-crh8b"] Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.646746 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-config-data\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.646832 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-fernet-keys\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.646859 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-credential-keys\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.646897 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-scripts\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.646939 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxjjv\" (UniqueName: \"kubernetes.io/projected/eaf4529b-f939-41eb-b31e-372ae95e1d51-kube-api-access-lxjjv\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.650678 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-scripts\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.651780 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-fernet-keys\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.652396 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-credential-keys\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.659850 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-config-data\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.661790 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxjjv\" (UniqueName: \"kubernetes.io/projected/eaf4529b-f939-41eb-b31e-372ae95e1d51-kube-api-access-lxjjv\") pod \"keystone-bootstrap-crh8b\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:10 crc kubenswrapper[4943]: I1204 10:35:10.853506 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:11 crc kubenswrapper[4943]: I1204 10:35:11.271630 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-crh8b"] Dec 04 10:35:11 crc kubenswrapper[4943]: I1204 10:35:11.355477 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" event={"ID":"eaf4529b-f939-41eb-b31e-372ae95e1d51","Type":"ContainerStarted","Data":"82227b6db3579163ee373022d704f7511f93c3da4d37d0122ecfcedf154a9964"} Dec 04 10:35:11 crc kubenswrapper[4943]: I1204 10:35:11.393557 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"osp-secret" Dec 04 10:35:12 crc kubenswrapper[4943]: I1204 10:35:12.362063 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" event={"ID":"eaf4529b-f939-41eb-b31e-372ae95e1d51","Type":"ContainerStarted","Data":"1e953ac8ed499eabcf2d34c691af5981551178fdb696ff7cba69f38327c886b1"} Dec 04 10:35:12 crc kubenswrapper[4943]: I1204 10:35:12.381351 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" podStartSLOduration=2.381332521 podStartE2EDuration="2.381332521s" podCreationTimestamp="2025-12-04 10:35:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:35:12.381154196 +0000 UTC m=+1160.970130064" watchObservedRunningTime="2025-12-04 10:35:12.381332521 +0000 UTC m=+1160.970308409" Dec 04 10:35:15 crc kubenswrapper[4943]: I1204 10:35:15.383974 4943 generic.go:334] "Generic (PLEG): container finished" podID="eaf4529b-f939-41eb-b31e-372ae95e1d51" containerID="1e953ac8ed499eabcf2d34c691af5981551178fdb696ff7cba69f38327c886b1" exitCode=0 Dec 04 10:35:15 crc kubenswrapper[4943]: I1204 10:35:15.384051 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" event={"ID":"eaf4529b-f939-41eb-b31e-372ae95e1d51","Type":"ContainerDied","Data":"1e953ac8ed499eabcf2d34c691af5981551178fdb696ff7cba69f38327c886b1"} Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.637404 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.833008 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-scripts\") pod \"eaf4529b-f939-41eb-b31e-372ae95e1d51\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.833074 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-credential-keys\") pod \"eaf4529b-f939-41eb-b31e-372ae95e1d51\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.833101 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-fernet-keys\") pod \"eaf4529b-f939-41eb-b31e-372ae95e1d51\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.833188 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-config-data\") pod \"eaf4529b-f939-41eb-b31e-372ae95e1d51\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.833252 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxjjv\" (UniqueName: \"kubernetes.io/projected/eaf4529b-f939-41eb-b31e-372ae95e1d51-kube-api-access-lxjjv\") pod \"eaf4529b-f939-41eb-b31e-372ae95e1d51\" (UID: \"eaf4529b-f939-41eb-b31e-372ae95e1d51\") " Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.842431 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "eaf4529b-f939-41eb-b31e-372ae95e1d51" (UID: "eaf4529b-f939-41eb-b31e-372ae95e1d51"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.842854 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-scripts" (OuterVolumeSpecName: "scripts") pod "eaf4529b-f939-41eb-b31e-372ae95e1d51" (UID: "eaf4529b-f939-41eb-b31e-372ae95e1d51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.843188 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaf4529b-f939-41eb-b31e-372ae95e1d51-kube-api-access-lxjjv" (OuterVolumeSpecName: "kube-api-access-lxjjv") pod "eaf4529b-f939-41eb-b31e-372ae95e1d51" (UID: "eaf4529b-f939-41eb-b31e-372ae95e1d51"). InnerVolumeSpecName "kube-api-access-lxjjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.851758 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-config-data" (OuterVolumeSpecName: "config-data") pod "eaf4529b-f939-41eb-b31e-372ae95e1d51" (UID: "eaf4529b-f939-41eb-b31e-372ae95e1d51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.854158 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "eaf4529b-f939-41eb-b31e-372ae95e1d51" (UID: "eaf4529b-f939-41eb-b31e-372ae95e1d51"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.935085 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.935124 4943 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.935135 4943 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.935144 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf4529b-f939-41eb-b31e-372ae95e1d51-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:16 crc kubenswrapper[4943]: I1204 10:35:16.935154 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxjjv\" (UniqueName: \"kubernetes.io/projected/eaf4529b-f939-41eb-b31e-372ae95e1d51-kube-api-access-lxjjv\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.410708 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.412573 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" event={"ID":"eaf4529b-f939-41eb-b31e-372ae95e1d51","Type":"ContainerDied","Data":"82227b6db3579163ee373022d704f7511f93c3da4d37d0122ecfcedf154a9964"} Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.412603 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82227b6db3579163ee373022d704f7511f93c3da4d37d0122ecfcedf154a9964" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.412652 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-crh8b" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.549390 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2"] Dec 04 10:35:17 crc kubenswrapper[4943]: E1204 10:35:17.549783 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf4529b-f939-41eb-b31e-372ae95e1d51" containerName="keystone-bootstrap" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.549807 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf4529b-f939-41eb-b31e-372ae95e1d51" containerName="keystone-bootstrap" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.549971 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf4529b-f939-41eb-b31e-372ae95e1d51" containerName="keystone-bootstrap" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.550578 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.553056 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-lc4gq" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.554268 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.554290 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.555651 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.621625 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2"] Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.745559 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-config-data\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.745610 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-fernet-keys\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.745638 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-scripts\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.745726 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-credential-keys\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.745774 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zvx8\" (UniqueName: \"kubernetes.io/projected/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-kube-api-access-4zvx8\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.846843 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-config-data\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.846892 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-fernet-keys\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.846914 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-scripts\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.846977 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-credential-keys\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.847023 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zvx8\" (UniqueName: \"kubernetes.io/projected/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-kube-api-access-4zvx8\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.852039 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-fernet-keys\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.853249 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-credential-keys\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.853359 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-config-data\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.857557 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-scripts\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.867661 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zvx8\" (UniqueName: \"kubernetes.io/projected/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-kube-api-access-4zvx8\") pod \"keystone-6d8bfb9775-vvbq2\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:17 crc kubenswrapper[4943]: I1204 10:35:17.868344 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:18 crc kubenswrapper[4943]: I1204 10:35:18.291870 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2"] Dec 04 10:35:18 crc kubenswrapper[4943]: I1204 10:35:18.420461 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" event={"ID":"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a","Type":"ContainerStarted","Data":"73e9654e5337260ddea4fab5e4e2306e99511f77d3bedca23ce24a1a5d503efd"} Dec 04 10:35:18 crc kubenswrapper[4943]: I1204 10:35:18.628986 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:35:18 crc kubenswrapper[4943]: I1204 10:35:18.629041 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:35:19 crc kubenswrapper[4943]: I1204 10:35:19.428789 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" event={"ID":"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a","Type":"ContainerStarted","Data":"5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf"} Dec 04 10:35:19 crc kubenswrapper[4943]: I1204 10:35:19.429222 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:19 crc kubenswrapper[4943]: I1204 10:35:19.453067 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" podStartSLOduration=2.453049119 podStartE2EDuration="2.453049119s" podCreationTimestamp="2025-12-04 10:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:35:19.448578805 +0000 UTC m=+1168.037554673" watchObservedRunningTime="2025-12-04 10:35:19.453049119 +0000 UTC m=+1168.042025007" Dec 04 10:35:20 crc kubenswrapper[4943]: I1204 10:35:20.982075 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-create-k85jz"] Dec 04 10:35:20 crc kubenswrapper[4943]: I1204 10:35:20.983058 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-k85jz" Dec 04 10:35:20 crc kubenswrapper[4943]: I1204 10:35:20.990416 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5"] Dec 04 10:35:20 crc kubenswrapper[4943]: I1204 10:35:20.991436 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" Dec 04 10:35:20 crc kubenswrapper[4943]: I1204 10:35:20.992922 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-db-secret" Dec 04 10:35:20 crc kubenswrapper[4943]: I1204 10:35:20.993620 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h485\" (UniqueName: \"kubernetes.io/projected/9bc2510f-d4d9-46dd-b5fc-409e89ababbd-kube-api-access-8h485\") pod \"cinder-db-create-k85jz\" (UID: \"9bc2510f-d4d9-46dd-b5fc-409e89ababbd\") " pod="cinder-kuttl-tests/cinder-db-create-k85jz" Dec 04 10:35:20 crc kubenswrapper[4943]: I1204 10:35:20.993690 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frttr\" (UniqueName: \"kubernetes.io/projected/121f8a1e-5d27-4f26-9578-65f06fd83846-kube-api-access-frttr\") pod \"cinder-2117-account-create-update-vzqx5\" (UID: \"121f8a1e-5d27-4f26-9578-65f06fd83846\") " pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" Dec 04 10:35:20 crc kubenswrapper[4943]: I1204 10:35:20.993899 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bc2510f-d4d9-46dd-b5fc-409e89ababbd-operator-scripts\") pod \"cinder-db-create-k85jz\" (UID: \"9bc2510f-d4d9-46dd-b5fc-409e89ababbd\") " pod="cinder-kuttl-tests/cinder-db-create-k85jz" Dec 04 10:35:20 crc kubenswrapper[4943]: I1204 10:35:20.993951 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/121f8a1e-5d27-4f26-9578-65f06fd83846-operator-scripts\") pod \"cinder-2117-account-create-update-vzqx5\" (UID: \"121f8a1e-5d27-4f26-9578-65f06fd83846\") " pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.001044 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-k85jz"] Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.006381 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5"] Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.110276 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bc2510f-d4d9-46dd-b5fc-409e89ababbd-operator-scripts\") pod \"cinder-db-create-k85jz\" (UID: \"9bc2510f-d4d9-46dd-b5fc-409e89ababbd\") " pod="cinder-kuttl-tests/cinder-db-create-k85jz" Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.110397 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/121f8a1e-5d27-4f26-9578-65f06fd83846-operator-scripts\") pod \"cinder-2117-account-create-update-vzqx5\" (UID: \"121f8a1e-5d27-4f26-9578-65f06fd83846\") " pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.110620 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h485\" (UniqueName: \"kubernetes.io/projected/9bc2510f-d4d9-46dd-b5fc-409e89ababbd-kube-api-access-8h485\") pod \"cinder-db-create-k85jz\" (UID: \"9bc2510f-d4d9-46dd-b5fc-409e89ababbd\") " pod="cinder-kuttl-tests/cinder-db-create-k85jz" Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.110682 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frttr\" (UniqueName: \"kubernetes.io/projected/121f8a1e-5d27-4f26-9578-65f06fd83846-kube-api-access-frttr\") pod \"cinder-2117-account-create-update-vzqx5\" (UID: \"121f8a1e-5d27-4f26-9578-65f06fd83846\") " pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.111379 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bc2510f-d4d9-46dd-b5fc-409e89ababbd-operator-scripts\") pod \"cinder-db-create-k85jz\" (UID: \"9bc2510f-d4d9-46dd-b5fc-409e89ababbd\") " pod="cinder-kuttl-tests/cinder-db-create-k85jz" Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.111401 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/121f8a1e-5d27-4f26-9578-65f06fd83846-operator-scripts\") pod \"cinder-2117-account-create-update-vzqx5\" (UID: \"121f8a1e-5d27-4f26-9578-65f06fd83846\") " pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.131177 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frttr\" (UniqueName: \"kubernetes.io/projected/121f8a1e-5d27-4f26-9578-65f06fd83846-kube-api-access-frttr\") pod \"cinder-2117-account-create-update-vzqx5\" (UID: \"121f8a1e-5d27-4f26-9578-65f06fd83846\") " pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.132288 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h485\" (UniqueName: \"kubernetes.io/projected/9bc2510f-d4d9-46dd-b5fc-409e89ababbd-kube-api-access-8h485\") pod \"cinder-db-create-k85jz\" (UID: \"9bc2510f-d4d9-46dd-b5fc-409e89ababbd\") " pod="cinder-kuttl-tests/cinder-db-create-k85jz" Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.306489 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-k85jz" Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.316317 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.712788 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-k85jz"] Dec 04 10:35:21 crc kubenswrapper[4943]: W1204 10:35:21.715943 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9bc2510f_d4d9_46dd_b5fc_409e89ababbd.slice/crio-7c1572f4726c28d1cc4af314b0d0fb0a2ae51cfb3c7e727f508fd896802c6165 WatchSource:0}: Error finding container 7c1572f4726c28d1cc4af314b0d0fb0a2ae51cfb3c7e727f508fd896802c6165: Status 404 returned error can't find the container with id 7c1572f4726c28d1cc4af314b0d0fb0a2ae51cfb3c7e727f508fd896802c6165 Dec 04 10:35:21 crc kubenswrapper[4943]: I1204 10:35:21.761274 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5"] Dec 04 10:35:21 crc kubenswrapper[4943]: W1204 10:35:21.770805 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod121f8a1e_5d27_4f26_9578_65f06fd83846.slice/crio-a547e6b3ae0cc839651ba2223d71985831fc08b87136cf5df0e63be635af249e WatchSource:0}: Error finding container a547e6b3ae0cc839651ba2223d71985831fc08b87136cf5df0e63be635af249e: Status 404 returned error can't find the container with id a547e6b3ae0cc839651ba2223d71985831fc08b87136cf5df0e63be635af249e Dec 04 10:35:22 crc kubenswrapper[4943]: I1204 10:35:22.453999 4943 generic.go:334] "Generic (PLEG): container finished" podID="9bc2510f-d4d9-46dd-b5fc-409e89ababbd" containerID="8d59862a25addea490182e432a477a56450b82e0ce71280a3fa7c5e66542d704" exitCode=0 Dec 04 10:35:22 crc kubenswrapper[4943]: I1204 10:35:22.454591 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-k85jz" event={"ID":"9bc2510f-d4d9-46dd-b5fc-409e89ababbd","Type":"ContainerDied","Data":"8d59862a25addea490182e432a477a56450b82e0ce71280a3fa7c5e66542d704"} Dec 04 10:35:22 crc kubenswrapper[4943]: I1204 10:35:22.454630 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-k85jz" event={"ID":"9bc2510f-d4d9-46dd-b5fc-409e89ababbd","Type":"ContainerStarted","Data":"7c1572f4726c28d1cc4af314b0d0fb0a2ae51cfb3c7e727f508fd896802c6165"} Dec 04 10:35:22 crc kubenswrapper[4943]: I1204 10:35:22.456587 4943 generic.go:334] "Generic (PLEG): container finished" podID="121f8a1e-5d27-4f26-9578-65f06fd83846" containerID="97d1e4bf9571948da1166e8ff207a37343c355e16f45d3d9629e3c6d1ee16256" exitCode=0 Dec 04 10:35:22 crc kubenswrapper[4943]: I1204 10:35:22.456620 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" event={"ID":"121f8a1e-5d27-4f26-9578-65f06fd83846","Type":"ContainerDied","Data":"97d1e4bf9571948da1166e8ff207a37343c355e16f45d3d9629e3c6d1ee16256"} Dec 04 10:35:22 crc kubenswrapper[4943]: I1204 10:35:22.456642 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" event={"ID":"121f8a1e-5d27-4f26-9578-65f06fd83846","Type":"ContainerStarted","Data":"a547e6b3ae0cc839651ba2223d71985831fc08b87136cf5df0e63be635af249e"} Dec 04 10:35:23 crc kubenswrapper[4943]: I1204 10:35:23.815851 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-k85jz" Dec 04 10:35:23 crc kubenswrapper[4943]: I1204 10:35:23.820085 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" Dec 04 10:35:23 crc kubenswrapper[4943]: I1204 10:35:23.960390 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/121f8a1e-5d27-4f26-9578-65f06fd83846-operator-scripts\") pod \"121f8a1e-5d27-4f26-9578-65f06fd83846\" (UID: \"121f8a1e-5d27-4f26-9578-65f06fd83846\") " Dec 04 10:35:23 crc kubenswrapper[4943]: I1204 10:35:23.960494 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frttr\" (UniqueName: \"kubernetes.io/projected/121f8a1e-5d27-4f26-9578-65f06fd83846-kube-api-access-frttr\") pod \"121f8a1e-5d27-4f26-9578-65f06fd83846\" (UID: \"121f8a1e-5d27-4f26-9578-65f06fd83846\") " Dec 04 10:35:23 crc kubenswrapper[4943]: I1204 10:35:23.960581 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bc2510f-d4d9-46dd-b5fc-409e89ababbd-operator-scripts\") pod \"9bc2510f-d4d9-46dd-b5fc-409e89ababbd\" (UID: \"9bc2510f-d4d9-46dd-b5fc-409e89ababbd\") " Dec 04 10:35:23 crc kubenswrapper[4943]: I1204 10:35:23.960661 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h485\" (UniqueName: \"kubernetes.io/projected/9bc2510f-d4d9-46dd-b5fc-409e89ababbd-kube-api-access-8h485\") pod \"9bc2510f-d4d9-46dd-b5fc-409e89ababbd\" (UID: \"9bc2510f-d4d9-46dd-b5fc-409e89ababbd\") " Dec 04 10:35:23 crc kubenswrapper[4943]: I1204 10:35:23.961464 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bc2510f-d4d9-46dd-b5fc-409e89ababbd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9bc2510f-d4d9-46dd-b5fc-409e89ababbd" (UID: "9bc2510f-d4d9-46dd-b5fc-409e89ababbd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:35:23 crc kubenswrapper[4943]: I1204 10:35:23.961495 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/121f8a1e-5d27-4f26-9578-65f06fd83846-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "121f8a1e-5d27-4f26-9578-65f06fd83846" (UID: "121f8a1e-5d27-4f26-9578-65f06fd83846"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:35:23 crc kubenswrapper[4943]: I1204 10:35:23.965579 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bc2510f-d4d9-46dd-b5fc-409e89ababbd-kube-api-access-8h485" (OuterVolumeSpecName: "kube-api-access-8h485") pod "9bc2510f-d4d9-46dd-b5fc-409e89ababbd" (UID: "9bc2510f-d4d9-46dd-b5fc-409e89ababbd"). InnerVolumeSpecName "kube-api-access-8h485". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:35:23 crc kubenswrapper[4943]: I1204 10:35:23.965902 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/121f8a1e-5d27-4f26-9578-65f06fd83846-kube-api-access-frttr" (OuterVolumeSpecName: "kube-api-access-frttr") pod "121f8a1e-5d27-4f26-9578-65f06fd83846" (UID: "121f8a1e-5d27-4f26-9578-65f06fd83846"). InnerVolumeSpecName "kube-api-access-frttr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:35:24 crc kubenswrapper[4943]: I1204 10:35:24.062617 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/121f8a1e-5d27-4f26-9578-65f06fd83846-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:24 crc kubenswrapper[4943]: I1204 10:35:24.062650 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frttr\" (UniqueName: \"kubernetes.io/projected/121f8a1e-5d27-4f26-9578-65f06fd83846-kube-api-access-frttr\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:24 crc kubenswrapper[4943]: I1204 10:35:24.062661 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9bc2510f-d4d9-46dd-b5fc-409e89ababbd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:24 crc kubenswrapper[4943]: I1204 10:35:24.062671 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h485\" (UniqueName: \"kubernetes.io/projected/9bc2510f-d4d9-46dd-b5fc-409e89ababbd-kube-api-access-8h485\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:24 crc kubenswrapper[4943]: I1204 10:35:24.471832 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" event={"ID":"121f8a1e-5d27-4f26-9578-65f06fd83846","Type":"ContainerDied","Data":"a547e6b3ae0cc839651ba2223d71985831fc08b87136cf5df0e63be635af249e"} Dec 04 10:35:24 crc kubenswrapper[4943]: I1204 10:35:24.471865 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5" Dec 04 10:35:24 crc kubenswrapper[4943]: I1204 10:35:24.471916 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a547e6b3ae0cc839651ba2223d71985831fc08b87136cf5df0e63be635af249e" Dec 04 10:35:24 crc kubenswrapper[4943]: I1204 10:35:24.473328 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-k85jz" event={"ID":"9bc2510f-d4d9-46dd-b5fc-409e89ababbd","Type":"ContainerDied","Data":"7c1572f4726c28d1cc4af314b0d0fb0a2ae51cfb3c7e727f508fd896802c6165"} Dec 04 10:35:24 crc kubenswrapper[4943]: I1204 10:35:24.473371 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c1572f4726c28d1cc4af314b0d0fb0a2ae51cfb3c7e727f508fd896802c6165" Dec 04 10:35:24 crc kubenswrapper[4943]: I1204 10:35:24.473374 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-k85jz" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.319992 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-jm9vm"] Dec 04 10:35:26 crc kubenswrapper[4943]: E1204 10:35:26.320762 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc2510f-d4d9-46dd-b5fc-409e89ababbd" containerName="mariadb-database-create" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.320794 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc2510f-d4d9-46dd-b5fc-409e89ababbd" containerName="mariadb-database-create" Dec 04 10:35:26 crc kubenswrapper[4943]: E1204 10:35:26.320812 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="121f8a1e-5d27-4f26-9578-65f06fd83846" containerName="mariadb-account-create-update" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.320818 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="121f8a1e-5d27-4f26-9578-65f06fd83846" containerName="mariadb-account-create-update" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.321001 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bc2510f-d4d9-46dd-b5fc-409e89ababbd" containerName="mariadb-database-create" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.321014 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="121f8a1e-5d27-4f26-9578-65f06fd83846" containerName="mariadb-account-create-update" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.321869 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.324160 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.324161 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-zlssz" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.324452 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.327965 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-jm9vm"] Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.497872 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-scripts\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.497957 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqqxn\" (UniqueName: \"kubernetes.io/projected/d3fd712b-e6d8-4282-9f14-9223a8e76bda-kube-api-access-lqqxn\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.498005 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3fd712b-e6d8-4282-9f14-9223a8e76bda-etc-machine-id\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.498038 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-db-sync-config-data\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.498066 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-config-data\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.599169 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqqxn\" (UniqueName: \"kubernetes.io/projected/d3fd712b-e6d8-4282-9f14-9223a8e76bda-kube-api-access-lqqxn\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.599517 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3fd712b-e6d8-4282-9f14-9223a8e76bda-etc-machine-id\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.599573 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-db-sync-config-data\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.599606 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-config-data\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.599712 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-scripts\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.599716 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3fd712b-e6d8-4282-9f14-9223a8e76bda-etc-machine-id\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.605563 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-config-data\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.605582 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-scripts\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.606280 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-db-sync-config-data\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.622221 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqqxn\" (UniqueName: \"kubernetes.io/projected/d3fd712b-e6d8-4282-9f14-9223a8e76bda-kube-api-access-lqqxn\") pod \"cinder-db-sync-jm9vm\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:26 crc kubenswrapper[4943]: I1204 10:35:26.669011 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:27 crc kubenswrapper[4943]: I1204 10:35:27.094502 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-jm9vm"] Dec 04 10:35:27 crc kubenswrapper[4943]: W1204 10:35:27.098484 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3fd712b_e6d8_4282_9f14_9223a8e76bda.slice/crio-3bd08e85def86d9b7d273737a17f402e0cc40a5a1a8004d7f10f400d4b0d6d4a WatchSource:0}: Error finding container 3bd08e85def86d9b7d273737a17f402e0cc40a5a1a8004d7f10f400d4b0d6d4a: Status 404 returned error can't find the container with id 3bd08e85def86d9b7d273737a17f402e0cc40a5a1a8004d7f10f400d4b0d6d4a Dec 04 10:35:27 crc kubenswrapper[4943]: I1204 10:35:27.493053 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" event={"ID":"d3fd712b-e6d8-4282-9f14-9223a8e76bda","Type":"ContainerStarted","Data":"3bd08e85def86d9b7d273737a17f402e0cc40a5a1a8004d7f10f400d4b0d6d4a"} Dec 04 10:35:42 crc kubenswrapper[4943]: I1204 10:35:42.620995 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" event={"ID":"d3fd712b-e6d8-4282-9f14-9223a8e76bda","Type":"ContainerStarted","Data":"b5b9f96b5feb0d6fb0d9f40cc14b1870320f8fefa81b6b1042864b78b72a1282"} Dec 04 10:35:42 crc kubenswrapper[4943]: I1204 10:35:42.647758 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" podStartSLOduration=2.608346139 podStartE2EDuration="16.647717818s" podCreationTimestamp="2025-12-04 10:35:26 +0000 UTC" firstStartedPulling="2025-12-04 10:35:27.1029299 +0000 UTC m=+1175.691905768" lastFinishedPulling="2025-12-04 10:35:41.142301569 +0000 UTC m=+1189.731277447" observedRunningTime="2025-12-04 10:35:42.647544603 +0000 UTC m=+1191.236520481" watchObservedRunningTime="2025-12-04 10:35:42.647717818 +0000 UTC m=+1191.236693706" Dec 04 10:35:47 crc kubenswrapper[4943]: I1204 10:35:47.653559 4943 generic.go:334] "Generic (PLEG): container finished" podID="d3fd712b-e6d8-4282-9f14-9223a8e76bda" containerID="b5b9f96b5feb0d6fb0d9f40cc14b1870320f8fefa81b6b1042864b78b72a1282" exitCode=0 Dec 04 10:35:47 crc kubenswrapper[4943]: I1204 10:35:47.653638 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" event={"ID":"d3fd712b-e6d8-4282-9f14-9223a8e76bda","Type":"ContainerDied","Data":"b5b9f96b5feb0d6fb0d9f40cc14b1870320f8fefa81b6b1042864b78b72a1282"} Dec 04 10:35:48 crc kubenswrapper[4943]: I1204 10:35:48.629408 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:35:48 crc kubenswrapper[4943]: I1204 10:35:48.629683 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:35:48 crc kubenswrapper[4943]: I1204 10:35:48.629752 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:35:48 crc kubenswrapper[4943]: I1204 10:35:48.630413 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"39ef936964dcabd0fd14176e3d0e5709046494c4af42c583afefd4c04502ae7a"} pod="openshift-machine-config-operator/machine-config-daemon-kswzd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 10:35:48 crc kubenswrapper[4943]: I1204 10:35:48.630524 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" containerID="cri-o://39ef936964dcabd0fd14176e3d0e5709046494c4af42c583afefd4c04502ae7a" gracePeriod=600 Dec 04 10:35:48 crc kubenswrapper[4943]: I1204 10:35:48.957783 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.149927 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-config-data\") pod \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.150014 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-db-sync-config-data\") pod \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.150050 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-scripts\") pod \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.150154 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqqxn\" (UniqueName: \"kubernetes.io/projected/d3fd712b-e6d8-4282-9f14-9223a8e76bda-kube-api-access-lqqxn\") pod \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.150237 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3fd712b-e6d8-4282-9f14-9223a8e76bda-etc-machine-id\") pod \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\" (UID: \"d3fd712b-e6d8-4282-9f14-9223a8e76bda\") " Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.150404 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3fd712b-e6d8-4282-9f14-9223a8e76bda-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d3fd712b-e6d8-4282-9f14-9223a8e76bda" (UID: "d3fd712b-e6d8-4282-9f14-9223a8e76bda"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.150620 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d3fd712b-e6d8-4282-9f14-9223a8e76bda-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.156020 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3fd712b-e6d8-4282-9f14-9223a8e76bda-kube-api-access-lqqxn" (OuterVolumeSpecName: "kube-api-access-lqqxn") pod "d3fd712b-e6d8-4282-9f14-9223a8e76bda" (UID: "d3fd712b-e6d8-4282-9f14-9223a8e76bda"). InnerVolumeSpecName "kube-api-access-lqqxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.156390 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d3fd712b-e6d8-4282-9f14-9223a8e76bda" (UID: "d3fd712b-e6d8-4282-9f14-9223a8e76bda"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.169139 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-scripts" (OuterVolumeSpecName: "scripts") pod "d3fd712b-e6d8-4282-9f14-9223a8e76bda" (UID: "d3fd712b-e6d8-4282-9f14-9223a8e76bda"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.199736 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-config-data" (OuterVolumeSpecName: "config-data") pod "d3fd712b-e6d8-4282-9f14-9223a8e76bda" (UID: "d3fd712b-e6d8-4282-9f14-9223a8e76bda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.251588 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqqxn\" (UniqueName: \"kubernetes.io/projected/d3fd712b-e6d8-4282-9f14-9223a8e76bda-kube-api-access-lqqxn\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.251621 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.251630 4943 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.251641 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3fd712b-e6d8-4282-9f14-9223a8e76bda-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.429953 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.668761 4943 generic.go:334] "Generic (PLEG): container finished" podID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerID="39ef936964dcabd0fd14176e3d0e5709046494c4af42c583afefd4c04502ae7a" exitCode=0 Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.668829 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerDied","Data":"39ef936964dcabd0fd14176e3d0e5709046494c4af42c583afefd4c04502ae7a"} Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.668876 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerStarted","Data":"7e58a73db13bfe564778718e69895f4217bb13940aaf3f0a781ae48e378dfd34"} Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.668929 4943 scope.go:117] "RemoveContainer" containerID="41363b73659bf0d33a8ed7f51e273d64bed37bf7e075e5c6bc173eb617215207" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.671602 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" event={"ID":"d3fd712b-e6d8-4282-9f14-9223a8e76bda","Type":"ContainerDied","Data":"3bd08e85def86d9b7d273737a17f402e0cc40a5a1a8004d7f10f400d4b0d6d4a"} Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.671688 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bd08e85def86d9b7d273737a17f402e0cc40a5a1a8004d7f10f400d4b0d6d4a" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.671807 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-jm9vm" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.940977 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 04 10:35:49 crc kubenswrapper[4943]: E1204 10:35:49.941291 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3fd712b-e6d8-4282-9f14-9223a8e76bda" containerName="cinder-db-sync" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.941308 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3fd712b-e6d8-4282-9f14-9223a8e76bda" containerName="cinder-db-sync" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.941419 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3fd712b-e6d8-4282-9f14-9223a8e76bda" containerName="cinder-db-sync" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.942020 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.943989 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.943993 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.944374 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-zlssz" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.944833 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scheduler-config-data" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.976229 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.978391 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.987091 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-backup-config-data" Dec 04 10:35:49 crc kubenswrapper[4943]: I1204 10:35:49.997852 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.007309 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.037426 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.038695 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.043747 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-volume-volume1-config-data" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.049956 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.070941 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.070984 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071006 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071021 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071035 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071048 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071067 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071087 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-config-data\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071105 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq6pd\" (UniqueName: \"kubernetes.io/projected/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-kube-api-access-kq6pd\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071125 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071305 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-scripts\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071358 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071386 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-sys\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071423 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071465 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071506 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-dev\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071534 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fglr\" (UniqueName: \"kubernetes.io/projected/dcc96db6-b608-422f-a293-689b490741c5-kube-api-access-7fglr\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071557 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071580 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-run\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071600 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-dev\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071624 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071679 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071713 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071758 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071778 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-sys\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071816 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-config-data-custom\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071853 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-lib-modules\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071887 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-scripts\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071907 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-run\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071937 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-config-data\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.071975 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.072006 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-nvme\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.072027 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snrz9\" (UniqueName: \"kubernetes.io/projected/5dbc6cc7-a807-4f74-ad16-feecbab07846-kube-api-access-snrz9\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.114477 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.115485 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.121034 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.130461 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.173993 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174051 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-sys\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174080 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-config-data-custom\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174112 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-config-data-custom\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174145 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-lib-modules\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174159 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174194 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-scripts\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174234 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-run\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174280 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-run\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174283 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-scripts\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174307 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-sys\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174336 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-config-data\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174443 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174477 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-nvme\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174501 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snrz9\" (UniqueName: \"kubernetes.io/projected/5dbc6cc7-a807-4f74-ad16-feecbab07846-kube-api-access-snrz9\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174527 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174579 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdrbm\" (UniqueName: \"kubernetes.io/projected/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-kube-api-access-fdrbm\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174631 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174678 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-config-data\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174711 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174738 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174764 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174786 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174816 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174854 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174890 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-config-data\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174918 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq6pd\" (UniqueName: \"kubernetes.io/projected/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-kube-api-access-kq6pd\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174938 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-logs\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.174972 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175007 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-scripts\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175033 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175059 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-sys\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175092 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175161 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175197 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-dev\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175213 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-lib-modules\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175247 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fglr\" (UniqueName: \"kubernetes.io/projected/dcc96db6-b608-422f-a293-689b490741c5-kube-api-access-7fglr\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175307 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175354 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-run\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175375 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-dev\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175401 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175451 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175477 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175811 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175869 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175904 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-run\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.175932 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-dev\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176035 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176121 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176257 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-nvme\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176260 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176429 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176532 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176586 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176854 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-dev\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176882 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176911 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176929 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176942 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-sys\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.176969 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.180881 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.182507 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-scripts\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.182653 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.182814 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-config-data-custom\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.182835 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-config-data\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.183367 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.183371 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-config-data\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.186415 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-scripts\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.192733 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.198784 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fglr\" (UniqueName: \"kubernetes.io/projected/dcc96db6-b608-422f-a293-689b490741c5-kube-api-access-7fglr\") pod \"cinder-backup-0\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.202378 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snrz9\" (UniqueName: \"kubernetes.io/projected/5dbc6cc7-a807-4f74-ad16-feecbab07846-kube-api-access-snrz9\") pod \"cinder-volume-volume1-0\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.203843 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq6pd\" (UniqueName: \"kubernetes.io/projected/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-kube-api-access-kq6pd\") pod \"cinder-scheduler-0\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.271578 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.276774 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-config-data\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.276818 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.276849 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-logs\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.276910 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-config-data-custom\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.276931 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-scripts\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.276958 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdrbm\" (UniqueName: \"kubernetes.io/projected/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-kube-api-access-fdrbm\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.277467 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.277694 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-logs\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.281053 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-config-data\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.281365 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-config-data-custom\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.281718 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-scripts\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.299436 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.301462 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdrbm\" (UniqueName: \"kubernetes.io/projected/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-kube-api-access-fdrbm\") pod \"cinder-api-0\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.362330 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.430766 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.646691 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.691131 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c","Type":"ContainerStarted","Data":"1c3a5773f69bf227bd2c5162dfc0fd97fb4b3dcc34f4cab0a6bff48b8f12d7bb"} Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.695662 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.716192 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 04 10:35:50 crc kubenswrapper[4943]: I1204 10:35:50.786905 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:35:50 crc kubenswrapper[4943]: W1204 10:35:50.793887 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8489a8d_2b17_4a37_b6f8_0da9b6a2318e.slice/crio-0a0a8f3e4c81bf96c99200a2dea237d6c5bf65bd62fa294b56be18a779500705 WatchSource:0}: Error finding container 0a0a8f3e4c81bf96c99200a2dea237d6c5bf65bd62fa294b56be18a779500705: Status 404 returned error can't find the container with id 0a0a8f3e4c81bf96c99200a2dea237d6c5bf65bd62fa294b56be18a779500705 Dec 04 10:35:51 crc kubenswrapper[4943]: I1204 10:35:51.699616 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e","Type":"ContainerStarted","Data":"7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf"} Dec 04 10:35:51 crc kubenswrapper[4943]: I1204 10:35:51.699987 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e","Type":"ContainerStarted","Data":"0a0a8f3e4c81bf96c99200a2dea237d6c5bf65bd62fa294b56be18a779500705"} Dec 04 10:35:51 crc kubenswrapper[4943]: I1204 10:35:51.701965 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"dcc96db6-b608-422f-a293-689b490741c5","Type":"ContainerStarted","Data":"2a71f9dcf8d9544e29935cf187f26cf0e141b45157ec9949221e1c7101ab4f04"} Dec 04 10:35:51 crc kubenswrapper[4943]: I1204 10:35:51.703103 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerStarted","Data":"b624f3c16d5991d677207b596e4e0beb3df23e66d031449746484c7fad83d725"} Dec 04 10:35:52 crc kubenswrapper[4943]: I1204 10:35:52.712244 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c","Type":"ContainerStarted","Data":"89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975"} Dec 04 10:35:52 crc kubenswrapper[4943]: I1204 10:35:52.714137 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e","Type":"ContainerStarted","Data":"32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1"} Dec 04 10:35:52 crc kubenswrapper[4943]: I1204 10:35:52.714288 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:35:52 crc kubenswrapper[4943]: I1204 10:35:52.719480 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"dcc96db6-b608-422f-a293-689b490741c5","Type":"ContainerStarted","Data":"ad98ac19f6f5811158d5e0064d662f1bd22e6182597088226f46659be1a2ead7"} Dec 04 10:35:52 crc kubenswrapper[4943]: I1204 10:35:52.719527 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"dcc96db6-b608-422f-a293-689b490741c5","Type":"ContainerStarted","Data":"30b3cdf6a5aed0a6e51a9d33ffd2692c4024394c53a586933ae10c02afd41e2d"} Dec 04 10:35:52 crc kubenswrapper[4943]: I1204 10:35:52.721488 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerStarted","Data":"59174cb53e5e85615e7704397d56fa30f8b671f005c21d0977bdd682584b259d"} Dec 04 10:35:52 crc kubenswrapper[4943]: I1204 10:35:52.721523 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerStarted","Data":"d1fa821e591f55221a38230037ff25ad7c1f08ea28329663f11848896499c40a"} Dec 04 10:35:52 crc kubenswrapper[4943]: I1204 10:35:52.739047 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=2.73813703 podStartE2EDuration="2.73813703s" podCreationTimestamp="2025-12-04 10:35:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:35:52.734759847 +0000 UTC m=+1201.323735715" watchObservedRunningTime="2025-12-04 10:35:52.73813703 +0000 UTC m=+1201.327112898" Dec 04 10:35:52 crc kubenswrapper[4943]: I1204 10:35:52.772223 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-0" podStartSLOduration=2.925271902 podStartE2EDuration="3.772184629s" podCreationTimestamp="2025-12-04 10:35:49 +0000 UTC" firstStartedPulling="2025-12-04 10:35:50.729179728 +0000 UTC m=+1199.318155596" lastFinishedPulling="2025-12-04 10:35:51.576092455 +0000 UTC m=+1200.165068323" observedRunningTime="2025-12-04 10:35:52.764788982 +0000 UTC m=+1201.353764860" watchObservedRunningTime="2025-12-04 10:35:52.772184629 +0000 UTC m=+1201.361160507" Dec 04 10:35:53 crc kubenswrapper[4943]: I1204 10:35:53.729752 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c","Type":"ContainerStarted","Data":"44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e"} Dec 04 10:35:53 crc kubenswrapper[4943]: I1204 10:35:53.751708 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podStartSLOduration=2.882935804 podStartE2EDuration="3.75168314s" podCreationTimestamp="2025-12-04 10:35:50 +0000 UTC" firstStartedPulling="2025-12-04 10:35:50.70736527 +0000 UTC m=+1199.296341138" lastFinishedPulling="2025-12-04 10:35:51.576112606 +0000 UTC m=+1200.165088474" observedRunningTime="2025-12-04 10:35:52.797932986 +0000 UTC m=+1201.386908854" watchObservedRunningTime="2025-12-04 10:35:53.75168314 +0000 UTC m=+1202.340658998" Dec 04 10:35:54 crc kubenswrapper[4943]: I1204 10:35:54.747431 4943 generic.go:334] "Generic (PLEG): container finished" podID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerID="59174cb53e5e85615e7704397d56fa30f8b671f005c21d0977bdd682584b259d" exitCode=1 Dec 04 10:35:54 crc kubenswrapper[4943]: I1204 10:35:54.748595 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerDied","Data":"59174cb53e5e85615e7704397d56fa30f8b671f005c21d0977bdd682584b259d"} Dec 04 10:35:54 crc kubenswrapper[4943]: I1204 10:35:54.748984 4943 scope.go:117] "RemoveContainer" containerID="59174cb53e5e85615e7704397d56fa30f8b671f005c21d0977bdd682584b259d" Dec 04 10:35:54 crc kubenswrapper[4943]: I1204 10:35:54.782401 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-0" podStartSLOduration=4.855239634 podStartE2EDuration="5.782382446s" podCreationTimestamp="2025-12-04 10:35:49 +0000 UTC" firstStartedPulling="2025-12-04 10:35:50.647448701 +0000 UTC m=+1199.236424569" lastFinishedPulling="2025-12-04 10:35:51.574591513 +0000 UTC m=+1200.163567381" observedRunningTime="2025-12-04 10:35:53.751676789 +0000 UTC m=+1202.340652667" watchObservedRunningTime="2025-12-04 10:35:54.782382446 +0000 UTC m=+1203.371358314" Dec 04 10:35:55 crc kubenswrapper[4943]: I1204 10:35:55.272392 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:35:55 crc kubenswrapper[4943]: I1204 10:35:55.299829 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:35:55 crc kubenswrapper[4943]: I1204 10:35:55.362605 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:55 crc kubenswrapper[4943]: I1204 10:35:55.757251 4943 generic.go:334] "Generic (PLEG): container finished" podID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerID="d1fa821e591f55221a38230037ff25ad7c1f08ea28329663f11848896499c40a" exitCode=1 Dec 04 10:35:55 crc kubenswrapper[4943]: I1204 10:35:55.758190 4943 scope.go:117] "RemoveContainer" containerID="d1fa821e591f55221a38230037ff25ad7c1f08ea28329663f11848896499c40a" Dec 04 10:35:55 crc kubenswrapper[4943]: I1204 10:35:55.758413 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerDied","Data":"d1fa821e591f55221a38230037ff25ad7c1f08ea28329663f11848896499c40a"} Dec 04 10:35:55 crc kubenswrapper[4943]: I1204 10:35:55.758441 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerStarted","Data":"e8c78a796a49e1126df0469078acae6ab6d10fd753ee0ea7edc6785b91f64cae"} Dec 04 10:35:56 crc kubenswrapper[4943]: I1204 10:35:56.363158 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:56 crc kubenswrapper[4943]: I1204 10:35:56.766185 4943 generic.go:334] "Generic (PLEG): container finished" podID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerID="e8c78a796a49e1126df0469078acae6ab6d10fd753ee0ea7edc6785b91f64cae" exitCode=1 Dec 04 10:35:56 crc kubenswrapper[4943]: I1204 10:35:56.766246 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerDied","Data":"e8c78a796a49e1126df0469078acae6ab6d10fd753ee0ea7edc6785b91f64cae"} Dec 04 10:35:56 crc kubenswrapper[4943]: I1204 10:35:56.766271 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerStarted","Data":"e34131fd2aae82d08491abaf84da26d66e8899e3895dd4ceb91083f0122ab85a"} Dec 04 10:35:56 crc kubenswrapper[4943]: I1204 10:35:56.766290 4943 scope.go:117] "RemoveContainer" containerID="59174cb53e5e85615e7704397d56fa30f8b671f005c21d0977bdd682584b259d" Dec 04 10:35:56 crc kubenswrapper[4943]: I1204 10:35:56.766988 4943 scope.go:117] "RemoveContainer" containerID="e8c78a796a49e1126df0469078acae6ab6d10fd753ee0ea7edc6785b91f64cae" Dec 04 10:35:56 crc kubenswrapper[4943]: E1204 10:35:56.767150 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:35:57 crc kubenswrapper[4943]: I1204 10:35:57.775186 4943 scope.go:117] "RemoveContainer" containerID="e8c78a796a49e1126df0469078acae6ab6d10fd753ee0ea7edc6785b91f64cae" Dec 04 10:35:57 crc kubenswrapper[4943]: E1204 10:35:57.775657 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:35:58 crc kubenswrapper[4943]: I1204 10:35:58.786831 4943 generic.go:334] "Generic (PLEG): container finished" podID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerID="e34131fd2aae82d08491abaf84da26d66e8899e3895dd4ceb91083f0122ab85a" exitCode=1 Dec 04 10:35:58 crc kubenswrapper[4943]: I1204 10:35:58.786901 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerDied","Data":"e34131fd2aae82d08491abaf84da26d66e8899e3895dd4ceb91083f0122ab85a"} Dec 04 10:35:58 crc kubenswrapper[4943]: I1204 10:35:58.787382 4943 scope.go:117] "RemoveContainer" containerID="d1fa821e591f55221a38230037ff25ad7c1f08ea28329663f11848896499c40a" Dec 04 10:35:58 crc kubenswrapper[4943]: I1204 10:35:58.787958 4943 scope.go:117] "RemoveContainer" containerID="e34131fd2aae82d08491abaf84da26d66e8899e3895dd4ceb91083f0122ab85a" Dec 04 10:35:58 crc kubenswrapper[4943]: I1204 10:35:58.787983 4943 scope.go:117] "RemoveContainer" containerID="e8c78a796a49e1126df0469078acae6ab6d10fd753ee0ea7edc6785b91f64cae" Dec 04 10:35:58 crc kubenswrapper[4943]: E1204 10:35:58.788222 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:35:59 crc kubenswrapper[4943]: I1204 10:35:59.363432 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:35:59 crc kubenswrapper[4943]: I1204 10:35:59.798347 4943 scope.go:117] "RemoveContainer" containerID="e34131fd2aae82d08491abaf84da26d66e8899e3895dd4ceb91083f0122ab85a" Dec 04 10:35:59 crc kubenswrapper[4943]: I1204 10:35:59.798750 4943 scope.go:117] "RemoveContainer" containerID="e8c78a796a49e1126df0469078acae6ab6d10fd753ee0ea7edc6785b91f64cae" Dec 04 10:35:59 crc kubenswrapper[4943]: E1204 10:35:59.799077 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:36:00 crc kubenswrapper[4943]: I1204 10:36:00.362708 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:36:00 crc kubenswrapper[4943]: I1204 10:36:00.362786 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:36:00 crc kubenswrapper[4943]: I1204 10:36:00.497614 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:36:00 crc kubenswrapper[4943]: I1204 10:36:00.512422 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:36:00 crc kubenswrapper[4943]: I1204 10:36:00.813435 4943 scope.go:117] "RemoveContainer" containerID="e34131fd2aae82d08491abaf84da26d66e8899e3895dd4ceb91083f0122ab85a" Dec 04 10:36:00 crc kubenswrapper[4943]: I1204 10:36:00.813468 4943 scope.go:117] "RemoveContainer" containerID="e8c78a796a49e1126df0469078acae6ab6d10fd753ee0ea7edc6785b91f64cae" Dec 04 10:36:00 crc kubenswrapper[4943]: E1204 10:36:00.813737 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:36:02 crc kubenswrapper[4943]: I1204 10:36:02.484487 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.653324 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.654443 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.665916 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.787922 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-scripts\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.787970 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70ec9572-3a43-4c57-9f86-badf121e54f4-etc-machine-id\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.788022 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-config-data-custom\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.788257 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qkd6\" (UniqueName: \"kubernetes.io/projected/70ec9572-3a43-4c57-9f86-badf121e54f4-kube-api-access-6qkd6\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.788318 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-config-data\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.889310 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-scripts\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.889579 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70ec9572-3a43-4c57-9f86-badf121e54f4-etc-machine-id\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.889703 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-config-data-custom\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.889823 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qkd6\" (UniqueName: \"kubernetes.io/projected/70ec9572-3a43-4c57-9f86-badf121e54f4-kube-api-access-6qkd6\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.889907 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-config-data\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.889707 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70ec9572-3a43-4c57-9f86-badf121e54f4-etc-machine-id\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.894549 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-config-data-custom\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.902594 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-scripts\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.903261 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-config-data\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.907277 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qkd6\" (UniqueName: \"kubernetes.io/projected/70ec9572-3a43-4c57-9f86-badf121e54f4-kube-api-access-6qkd6\") pod \"cinder-scheduler-1\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:03 crc kubenswrapper[4943]: I1204 10:36:03.972831 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:04 crc kubenswrapper[4943]: W1204 10:36:04.381602 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70ec9572_3a43_4c57_9f86_badf121e54f4.slice/crio-3aaad3521c72eaa961366de803fb76bd4216f30561b29fe101f5b1764d7dbe70 WatchSource:0}: Error finding container 3aaad3521c72eaa961366de803fb76bd4216f30561b29fe101f5b1764d7dbe70: Status 404 returned error can't find the container with id 3aaad3521c72eaa961366de803fb76bd4216f30561b29fe101f5b1764d7dbe70 Dec 04 10:36:04 crc kubenswrapper[4943]: I1204 10:36:04.385022 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 04 10:36:04 crc kubenswrapper[4943]: I1204 10:36:04.836789 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"70ec9572-3a43-4c57-9f86-badf121e54f4","Type":"ContainerStarted","Data":"3aaad3521c72eaa961366de803fb76bd4216f30561b29fe101f5b1764d7dbe70"} Dec 04 10:36:05 crc kubenswrapper[4943]: I1204 10:36:05.845850 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"70ec9572-3a43-4c57-9f86-badf121e54f4","Type":"ContainerStarted","Data":"154a16d6697e732511345ef9aa83b0b87032dafcab7e52b614202fff2e49f8df"} Dec 04 10:36:06 crc kubenswrapper[4943]: I1204 10:36:06.854755 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"70ec9572-3a43-4c57-9f86-badf121e54f4","Type":"ContainerStarted","Data":"6d95b46a9e65d3fd1f7881d0d45edd49cb14544c6fc5d3563720208f62178960"} Dec 04 10:36:08 crc kubenswrapper[4943]: I1204 10:36:08.973102 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:13 crc kubenswrapper[4943]: I1204 10:36:13.565869 4943 scope.go:117] "RemoveContainer" containerID="e34131fd2aae82d08491abaf84da26d66e8899e3895dd4ceb91083f0122ab85a" Dec 04 10:36:13 crc kubenswrapper[4943]: I1204 10:36:13.565933 4943 scope.go:117] "RemoveContainer" containerID="e8c78a796a49e1126df0469078acae6ab6d10fd753ee0ea7edc6785b91f64cae" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.153020 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.168968 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-1" podStartSLOduration=11.168951403 podStartE2EDuration="11.168951403s" podCreationTimestamp="2025-12-04 10:36:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:36:06.88095096 +0000 UTC m=+1215.469926828" watchObservedRunningTime="2025-12-04 10:36:14.168951403 +0000 UTC m=+1222.757927271" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.215928 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.217892 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.227727 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.276959 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-config-data\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.277283 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-config-data-custom\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.277465 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s84pm\" (UniqueName: \"kubernetes.io/projected/4e33289e-7d09-40e5-a914-813a62486644-kube-api-access-s84pm\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.277612 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-scripts\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.277749 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e33289e-7d09-40e5-a914-813a62486644-etc-machine-id\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.379441 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-config-data\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.379496 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-config-data-custom\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.379564 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s84pm\" (UniqueName: \"kubernetes.io/projected/4e33289e-7d09-40e5-a914-813a62486644-kube-api-access-s84pm\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.379596 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-scripts\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.379622 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e33289e-7d09-40e5-a914-813a62486644-etc-machine-id\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.379711 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e33289e-7d09-40e5-a914-813a62486644-etc-machine-id\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.384957 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-config-data-custom\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.385483 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-scripts\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.387184 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-config-data\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.396594 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s84pm\" (UniqueName: \"kubernetes.io/projected/4e33289e-7d09-40e5-a914-813a62486644-kube-api-access-s84pm\") pod \"cinder-scheduler-2\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.546065 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:14 crc kubenswrapper[4943]: I1204 10:36:14.976750 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 04 10:36:14 crc kubenswrapper[4943]: W1204 10:36:14.981392 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e33289e_7d09_40e5_a914_813a62486644.slice/crio-a8fd72ba38688b9977059cf97ce95bb5277eef53e67aedd244d51c0e0f7d96b3 WatchSource:0}: Error finding container a8fd72ba38688b9977059cf97ce95bb5277eef53e67aedd244d51c0e0f7d96b3: Status 404 returned error can't find the container with id a8fd72ba38688b9977059cf97ce95bb5277eef53e67aedd244d51c0e0f7d96b3 Dec 04 10:36:15 crc kubenswrapper[4943]: I1204 10:36:15.931981 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"4e33289e-7d09-40e5-a914-813a62486644","Type":"ContainerStarted","Data":"a8fd72ba38688b9977059cf97ce95bb5277eef53e67aedd244d51c0e0f7d96b3"} Dec 04 10:36:16 crc kubenswrapper[4943]: I1204 10:36:16.940126 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerStarted","Data":"9dbe4b564c22dcdc28c5907566e1deee1d2273994bc376456b1a2f567a7d4255"} Dec 04 10:36:16 crc kubenswrapper[4943]: I1204 10:36:16.940642 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerStarted","Data":"773d56d8459293541800c4ad4e06c3d84b13851e31fdf7aaa2de8e84e38749e8"} Dec 04 10:36:16 crc kubenswrapper[4943]: I1204 10:36:16.945360 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"4e33289e-7d09-40e5-a914-813a62486644","Type":"ContainerStarted","Data":"e9505fadeeef9a149f4de9857cc3b2147be85913d3337bd371aa5bcb00a93ad6"} Dec 04 10:36:16 crc kubenswrapper[4943]: I1204 10:36:16.945401 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"4e33289e-7d09-40e5-a914-813a62486644","Type":"ContainerStarted","Data":"faae4c3c8e2dfae8df21fb63ea9c7a63f95020aa06bd029b7f332791d7c5ca24"} Dec 04 10:36:16 crc kubenswrapper[4943]: I1204 10:36:16.985715 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-2" podStartSLOduration=2.9856981339999997 podStartE2EDuration="2.985698134s" podCreationTimestamp="2025-12-04 10:36:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:36:16.981249591 +0000 UTC m=+1225.570225469" watchObservedRunningTime="2025-12-04 10:36:16.985698134 +0000 UTC m=+1225.574674002" Dec 04 10:36:18 crc kubenswrapper[4943]: I1204 10:36:18.976742 4943 generic.go:334] "Generic (PLEG): container finished" podID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerID="9dbe4b564c22dcdc28c5907566e1deee1d2273994bc376456b1a2f567a7d4255" exitCode=1 Dec 04 10:36:18 crc kubenswrapper[4943]: I1204 10:36:18.976938 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerDied","Data":"9dbe4b564c22dcdc28c5907566e1deee1d2273994bc376456b1a2f567a7d4255"} Dec 04 10:36:18 crc kubenswrapper[4943]: I1204 10:36:18.977412 4943 scope.go:117] "RemoveContainer" containerID="e8c78a796a49e1126df0469078acae6ab6d10fd753ee0ea7edc6785b91f64cae" Dec 04 10:36:18 crc kubenswrapper[4943]: I1204 10:36:18.978155 4943 scope.go:117] "RemoveContainer" containerID="9dbe4b564c22dcdc28c5907566e1deee1d2273994bc376456b1a2f567a7d4255" Dec 04 10:36:18 crc kubenswrapper[4943]: E1204 10:36:18.978494 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:36:19 crc kubenswrapper[4943]: I1204 10:36:19.546771 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:19 crc kubenswrapper[4943]: I1204 10:36:19.994058 4943 generic.go:334] "Generic (PLEG): container finished" podID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerID="773d56d8459293541800c4ad4e06c3d84b13851e31fdf7aaa2de8e84e38749e8" exitCode=1 Dec 04 10:36:19 crc kubenswrapper[4943]: I1204 10:36:19.994127 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerDied","Data":"773d56d8459293541800c4ad4e06c3d84b13851e31fdf7aaa2de8e84e38749e8"} Dec 04 10:36:19 crc kubenswrapper[4943]: I1204 10:36:19.994219 4943 scope.go:117] "RemoveContainer" containerID="e34131fd2aae82d08491abaf84da26d66e8899e3895dd4ceb91083f0122ab85a" Dec 04 10:36:19 crc kubenswrapper[4943]: I1204 10:36:19.995039 4943 scope.go:117] "RemoveContainer" containerID="773d56d8459293541800c4ad4e06c3d84b13851e31fdf7aaa2de8e84e38749e8" Dec 04 10:36:19 crc kubenswrapper[4943]: I1204 10:36:19.995141 4943 scope.go:117] "RemoveContainer" containerID="9dbe4b564c22dcdc28c5907566e1deee1d2273994bc376456b1a2f567a7d4255" Dec 04 10:36:19 crc kubenswrapper[4943]: E1204 10:36:19.995509 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:36:20 crc kubenswrapper[4943]: I1204 10:36:20.362500 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:36:20 crc kubenswrapper[4943]: I1204 10:36:20.362820 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:36:20 crc kubenswrapper[4943]: I1204 10:36:20.362838 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:36:21 crc kubenswrapper[4943]: I1204 10:36:21.005111 4943 scope.go:117] "RemoveContainer" containerID="773d56d8459293541800c4ad4e06c3d84b13851e31fdf7aaa2de8e84e38749e8" Dec 04 10:36:21 crc kubenswrapper[4943]: I1204 10:36:21.005144 4943 scope.go:117] "RemoveContainer" containerID="9dbe4b564c22dcdc28c5907566e1deee1d2273994bc376456b1a2f567a7d4255" Dec 04 10:36:21 crc kubenswrapper[4943]: E1204 10:36:21.005417 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:36:24 crc kubenswrapper[4943]: I1204 10:36:24.732735 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:25 crc kubenswrapper[4943]: I1204 10:36:25.165411 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 04 10:36:25 crc kubenswrapper[4943]: I1204 10:36:25.165916 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-2" podUID="4e33289e-7d09-40e5-a914-813a62486644" containerName="cinder-scheduler" containerID="cri-o://faae4c3c8e2dfae8df21fb63ea9c7a63f95020aa06bd029b7f332791d7c5ca24" gracePeriod=30 Dec 04 10:36:25 crc kubenswrapper[4943]: I1204 10:36:25.166078 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-2" podUID="4e33289e-7d09-40e5-a914-813a62486644" containerName="probe" containerID="cri-o://e9505fadeeef9a149f4de9857cc3b2147be85913d3337bd371aa5bcb00a93ad6" gracePeriod=30 Dec 04 10:36:26 crc kubenswrapper[4943]: I1204 10:36:26.046753 4943 generic.go:334] "Generic (PLEG): container finished" podID="4e33289e-7d09-40e5-a914-813a62486644" containerID="e9505fadeeef9a149f4de9857cc3b2147be85913d3337bd371aa5bcb00a93ad6" exitCode=0 Dec 04 10:36:26 crc kubenswrapper[4943]: I1204 10:36:26.046799 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"4e33289e-7d09-40e5-a914-813a62486644","Type":"ContainerDied","Data":"e9505fadeeef9a149f4de9857cc3b2147be85913d3337bd371aa5bcb00a93ad6"} Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.056990 4943 generic.go:334] "Generic (PLEG): container finished" podID="4e33289e-7d09-40e5-a914-813a62486644" containerID="faae4c3c8e2dfae8df21fb63ea9c7a63f95020aa06bd029b7f332791d7c5ca24" exitCode=0 Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.057423 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"4e33289e-7d09-40e5-a914-813a62486644","Type":"ContainerDied","Data":"faae4c3c8e2dfae8df21fb63ea9c7a63f95020aa06bd029b7f332791d7c5ca24"} Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.251432 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.287754 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s84pm\" (UniqueName: \"kubernetes.io/projected/4e33289e-7d09-40e5-a914-813a62486644-kube-api-access-s84pm\") pod \"4e33289e-7d09-40e5-a914-813a62486644\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.288007 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-config-data\") pod \"4e33289e-7d09-40e5-a914-813a62486644\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.288028 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-scripts\") pod \"4e33289e-7d09-40e5-a914-813a62486644\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.288053 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e33289e-7d09-40e5-a914-813a62486644-etc-machine-id\") pod \"4e33289e-7d09-40e5-a914-813a62486644\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.288076 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-config-data-custom\") pod \"4e33289e-7d09-40e5-a914-813a62486644\" (UID: \"4e33289e-7d09-40e5-a914-813a62486644\") " Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.290516 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4e33289e-7d09-40e5-a914-813a62486644-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4e33289e-7d09-40e5-a914-813a62486644" (UID: "4e33289e-7d09-40e5-a914-813a62486644"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.296068 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-scripts" (OuterVolumeSpecName: "scripts") pod "4e33289e-7d09-40e5-a914-813a62486644" (UID: "4e33289e-7d09-40e5-a914-813a62486644"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.297353 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4e33289e-7d09-40e5-a914-813a62486644" (UID: "4e33289e-7d09-40e5-a914-813a62486644"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.297607 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e33289e-7d09-40e5-a914-813a62486644-kube-api-access-s84pm" (OuterVolumeSpecName: "kube-api-access-s84pm") pod "4e33289e-7d09-40e5-a914-813a62486644" (UID: "4e33289e-7d09-40e5-a914-813a62486644"). InnerVolumeSpecName "kube-api-access-s84pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.360990 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-config-data" (OuterVolumeSpecName: "config-data") pod "4e33289e-7d09-40e5-a914-813a62486644" (UID: "4e33289e-7d09-40e5-a914-813a62486644"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.390446 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.390483 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.390493 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4e33289e-7d09-40e5-a914-813a62486644-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.390503 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4e33289e-7d09-40e5-a914-813a62486644-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:27 crc kubenswrapper[4943]: I1204 10:36:27.390514 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s84pm\" (UniqueName: \"kubernetes.io/projected/4e33289e-7d09-40e5-a914-813a62486644-kube-api-access-s84pm\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:28 crc kubenswrapper[4943]: I1204 10:36:28.066876 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-2" event={"ID":"4e33289e-7d09-40e5-a914-813a62486644","Type":"ContainerDied","Data":"a8fd72ba38688b9977059cf97ce95bb5277eef53e67aedd244d51c0e0f7d96b3"} Dec 04 10:36:28 crc kubenswrapper[4943]: I1204 10:36:28.067248 4943 scope.go:117] "RemoveContainer" containerID="e9505fadeeef9a149f4de9857cc3b2147be85913d3337bd371aa5bcb00a93ad6" Dec 04 10:36:28 crc kubenswrapper[4943]: I1204 10:36:28.066926 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-2" Dec 04 10:36:28 crc kubenswrapper[4943]: I1204 10:36:28.093111 4943 scope.go:117] "RemoveContainer" containerID="faae4c3c8e2dfae8df21fb63ea9c7a63f95020aa06bd029b7f332791d7c5ca24" Dec 04 10:36:28 crc kubenswrapper[4943]: I1204 10:36:28.099299 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 04 10:36:28 crc kubenswrapper[4943]: I1204 10:36:28.115183 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-2"] Dec 04 10:36:28 crc kubenswrapper[4943]: I1204 10:36:28.121004 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 04 10:36:28 crc kubenswrapper[4943]: I1204 10:36:28.121286 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-1" podUID="70ec9572-3a43-4c57-9f86-badf121e54f4" containerName="cinder-scheduler" containerID="cri-o://154a16d6697e732511345ef9aa83b0b87032dafcab7e52b614202fff2e49f8df" gracePeriod=30 Dec 04 10:36:28 crc kubenswrapper[4943]: I1204 10:36:28.121347 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-1" podUID="70ec9572-3a43-4c57-9f86-badf121e54f4" containerName="probe" containerID="cri-o://6d95b46a9e65d3fd1f7881d0d45edd49cb14544c6fc5d3563720208f62178960" gracePeriod=30 Dec 04 10:36:28 crc kubenswrapper[4943]: I1204 10:36:28.572435 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e33289e-7d09-40e5-a914-813a62486644" path="/var/lib/kubelet/pods/4e33289e-7d09-40e5-a914-813a62486644/volumes" Dec 04 10:36:29 crc kubenswrapper[4943]: I1204 10:36:29.078002 4943 generic.go:334] "Generic (PLEG): container finished" podID="70ec9572-3a43-4c57-9f86-badf121e54f4" containerID="6d95b46a9e65d3fd1f7881d0d45edd49cb14544c6fc5d3563720208f62178960" exitCode=0 Dec 04 10:36:29 crc kubenswrapper[4943]: I1204 10:36:29.078126 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"70ec9572-3a43-4c57-9f86-badf121e54f4","Type":"ContainerDied","Data":"6d95b46a9e65d3fd1f7881d0d45edd49cb14544c6fc5d3563720208f62178960"} Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.103548 4943 generic.go:334] "Generic (PLEG): container finished" podID="70ec9572-3a43-4c57-9f86-badf121e54f4" containerID="154a16d6697e732511345ef9aa83b0b87032dafcab7e52b614202fff2e49f8df" exitCode=0 Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.103602 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"70ec9572-3a43-4c57-9f86-badf121e54f4","Type":"ContainerDied","Data":"154a16d6697e732511345ef9aa83b0b87032dafcab7e52b614202fff2e49f8df"} Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.202948 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.284411 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70ec9572-3a43-4c57-9f86-badf121e54f4-etc-machine-id\") pod \"70ec9572-3a43-4c57-9f86-badf121e54f4\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.284475 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qkd6\" (UniqueName: \"kubernetes.io/projected/70ec9572-3a43-4c57-9f86-badf121e54f4-kube-api-access-6qkd6\") pod \"70ec9572-3a43-4c57-9f86-badf121e54f4\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.284598 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-scripts\") pod \"70ec9572-3a43-4c57-9f86-badf121e54f4\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.284652 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-config-data-custom\") pod \"70ec9572-3a43-4c57-9f86-badf121e54f4\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.284734 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-config-data\") pod \"70ec9572-3a43-4c57-9f86-badf121e54f4\" (UID: \"70ec9572-3a43-4c57-9f86-badf121e54f4\") " Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.285966 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70ec9572-3a43-4c57-9f86-badf121e54f4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "70ec9572-3a43-4c57-9f86-badf121e54f4" (UID: "70ec9572-3a43-4c57-9f86-badf121e54f4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.291172 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70ec9572-3a43-4c57-9f86-badf121e54f4-kube-api-access-6qkd6" (OuterVolumeSpecName: "kube-api-access-6qkd6") pod "70ec9572-3a43-4c57-9f86-badf121e54f4" (UID: "70ec9572-3a43-4c57-9f86-badf121e54f4"). InnerVolumeSpecName "kube-api-access-6qkd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.297447 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-scripts" (OuterVolumeSpecName: "scripts") pod "70ec9572-3a43-4c57-9f86-badf121e54f4" (UID: "70ec9572-3a43-4c57-9f86-badf121e54f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.306582 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "70ec9572-3a43-4c57-9f86-badf121e54f4" (UID: "70ec9572-3a43-4c57-9f86-badf121e54f4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.365826 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-config-data" (OuterVolumeSpecName: "config-data") pod "70ec9572-3a43-4c57-9f86-badf121e54f4" (UID: "70ec9572-3a43-4c57-9f86-badf121e54f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.385941 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.386184 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.386331 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70ec9572-3a43-4c57-9f86-badf121e54f4-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.386420 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70ec9572-3a43-4c57-9f86-badf121e54f4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:32 crc kubenswrapper[4943]: I1204 10:36:32.386505 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qkd6\" (UniqueName: \"kubernetes.io/projected/70ec9572-3a43-4c57-9f86-badf121e54f4-kube-api-access-6qkd6\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.114614 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-1" event={"ID":"70ec9572-3a43-4c57-9f86-badf121e54f4","Type":"ContainerDied","Data":"3aaad3521c72eaa961366de803fb76bd4216f30561b29fe101f5b1764d7dbe70"} Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.114684 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.115152 4943 scope.go:117] "RemoveContainer" containerID="6d95b46a9e65d3fd1f7881d0d45edd49cb14544c6fc5d3563720208f62178960" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.149184 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.150575 4943 scope.go:117] "RemoveContainer" containerID="154a16d6697e732511345ef9aa83b0b87032dafcab7e52b614202fff2e49f8df" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.156549 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-1"] Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.535398 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 04 10:36:33 crc kubenswrapper[4943]: E1204 10:36:33.535742 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ec9572-3a43-4c57-9f86-badf121e54f4" containerName="probe" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.535763 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ec9572-3a43-4c57-9f86-badf121e54f4" containerName="probe" Dec 04 10:36:33 crc kubenswrapper[4943]: E1204 10:36:33.535784 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70ec9572-3a43-4c57-9f86-badf121e54f4" containerName="cinder-scheduler" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.535793 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="70ec9572-3a43-4c57-9f86-badf121e54f4" containerName="cinder-scheduler" Dec 04 10:36:33 crc kubenswrapper[4943]: E1204 10:36:33.535812 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e33289e-7d09-40e5-a914-813a62486644" containerName="cinder-scheduler" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.535820 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e33289e-7d09-40e5-a914-813a62486644" containerName="cinder-scheduler" Dec 04 10:36:33 crc kubenswrapper[4943]: E1204 10:36:33.535830 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e33289e-7d09-40e5-a914-813a62486644" containerName="probe" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.535838 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e33289e-7d09-40e5-a914-813a62486644" containerName="probe" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.536004 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e33289e-7d09-40e5-a914-813a62486644" containerName="probe" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.536020 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="70ec9572-3a43-4c57-9f86-badf121e54f4" containerName="cinder-scheduler" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.536032 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e33289e-7d09-40e5-a914-813a62486644" containerName="cinder-scheduler" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.536044 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="70ec9572-3a43-4c57-9f86-badf121e54f4" containerName="probe" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.536945 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.545719 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.565179 4943 scope.go:117] "RemoveContainer" containerID="773d56d8459293541800c4ad4e06c3d84b13851e31fdf7aaa2de8e84e38749e8" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.565218 4943 scope.go:117] "RemoveContainer" containerID="9dbe4b564c22dcdc28c5907566e1deee1d2273994bc376456b1a2f567a7d4255" Dec 04 10:36:33 crc kubenswrapper[4943]: E1204 10:36:33.565433 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603001 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-lib-cinder\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603052 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-machine-id\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603085 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2265w\" (UniqueName: \"kubernetes.io/projected/f44de5e7-3753-4159-b01c-970889c7adf7-kube-api-access-2265w\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603169 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-sys\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603227 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-lib-modules\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603253 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-config-data\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603276 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-iscsi\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603293 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-scripts\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603343 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-nvme\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603371 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-dev\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603398 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-locks-cinder\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603451 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-run\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603507 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-locks-brick\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.603565 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-config-data-custom\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.704682 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-sys\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.704731 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-lib-modules\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.704763 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-config-data\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.704784 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-iscsi\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.704800 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-scripts\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.704838 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-nvme\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.704876 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-dev\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.704893 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-locks-cinder\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.704909 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-run\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.704885 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-lib-modules\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.704939 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-locks-brick\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.705146 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-config-data-custom\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.705260 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-lib-cinder\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.705291 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-machine-id\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.705360 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2265w\" (UniqueName: \"kubernetes.io/projected/f44de5e7-3753-4159-b01c-970889c7adf7-kube-api-access-2265w\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.705534 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-lib-cinder\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.705535 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-machine-id\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.705653 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-run\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.705862 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-dev\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.705922 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-locks-cinder\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.705927 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-locks-brick\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.705964 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-iscsi\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.706037 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-sys\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.706122 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-nvme\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.709804 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-config-data-custom\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.710681 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-scripts\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.711070 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-config-data\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.724282 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2265w\" (UniqueName: \"kubernetes.io/projected/f44de5e7-3753-4159-b01c-970889c7adf7-kube-api-access-2265w\") pod \"cinder-backup-1\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:33 crc kubenswrapper[4943]: I1204 10:36:33.851996 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:34 crc kubenswrapper[4943]: I1204 10:36:34.261322 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 04 10:36:34 crc kubenswrapper[4943]: W1204 10:36:34.265750 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf44de5e7_3753_4159_b01c_970889c7adf7.slice/crio-c41bb7a39cadc1137089c6f0c38c3d1cffbf0c5549bc4bfdc8aeab37badc99ae WatchSource:0}: Error finding container c41bb7a39cadc1137089c6f0c38c3d1cffbf0c5549bc4bfdc8aeab37badc99ae: Status 404 returned error can't find the container with id c41bb7a39cadc1137089c6f0c38c3d1cffbf0c5549bc4bfdc8aeab37badc99ae Dec 04 10:36:34 crc kubenswrapper[4943]: I1204 10:36:34.571572 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70ec9572-3a43-4c57-9f86-badf121e54f4" path="/var/lib/kubelet/pods/70ec9572-3a43-4c57-9f86-badf121e54f4/volumes" Dec 04 10:36:35 crc kubenswrapper[4943]: I1204 10:36:35.132408 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"f44de5e7-3753-4159-b01c-970889c7adf7","Type":"ContainerStarted","Data":"95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441"} Dec 04 10:36:35 crc kubenswrapper[4943]: I1204 10:36:35.132704 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"f44de5e7-3753-4159-b01c-970889c7adf7","Type":"ContainerStarted","Data":"f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969"} Dec 04 10:36:35 crc kubenswrapper[4943]: I1204 10:36:35.132808 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"f44de5e7-3753-4159-b01c-970889c7adf7","Type":"ContainerStarted","Data":"c41bb7a39cadc1137089c6f0c38c3d1cffbf0c5549bc4bfdc8aeab37badc99ae"} Dec 04 10:36:35 crc kubenswrapper[4943]: I1204 10:36:35.155538 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-1" podStartSLOduration=2.155513022 podStartE2EDuration="2.155513022s" podCreationTimestamp="2025-12-04 10:36:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:36:35.152595212 +0000 UTC m=+1243.741571100" watchObservedRunningTime="2025-12-04 10:36:35.155513022 +0000 UTC m=+1243.744488890" Dec 04 10:36:38 crc kubenswrapper[4943]: I1204 10:36:38.852852 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.024226 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.101302 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.102383 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.114626 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.160028 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-lib-cinder\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.160091 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-nvme\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.160121 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-locks-brick\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.160146 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-dev\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.160182 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-sys\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.160229 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-iscsi\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.160259 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-config-data\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.160584 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-config-data-custom\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.160688 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-scripts\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.160984 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-machine-id\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.161078 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-locks-cinder\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.161172 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-run\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.161271 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-lib-modules\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.161325 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl6bc\" (UniqueName: \"kubernetes.io/projected/20dd5c87-1885-44c6-8ed4-97898e603817-kube-api-access-rl6bc\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263073 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-sys\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263128 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-iscsi\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263157 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-config-data\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263224 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-config-data-custom\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263223 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-sys\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263254 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-scripts\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263350 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-machine-id\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263397 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-iscsi\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263521 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-machine-id\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263571 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-locks-cinder\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263616 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-run\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263630 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-locks-cinder\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263646 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-lib-modules\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263679 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-lib-modules\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263718 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl6bc\" (UniqueName: \"kubernetes.io/projected/20dd5c87-1885-44c6-8ed4-97898e603817-kube-api-access-rl6bc\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263749 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-lib-cinder\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263777 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-nvme\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263802 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-locks-brick\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263720 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-run\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263920 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-dev\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263962 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-lib-cinder\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263992 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-dev\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263988 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-locks-brick\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.263896 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-nvme\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.269714 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-config-data-custom\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.271646 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-scripts\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.273683 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-config-data\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.287173 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl6bc\" (UniqueName: \"kubernetes.io/projected/20dd5c87-1885-44c6-8ed4-97898e603817-kube-api-access-rl6bc\") pod \"cinder-backup-2\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.421675 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:44 crc kubenswrapper[4943]: I1204 10:36:44.845743 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 04 10:36:45 crc kubenswrapper[4943]: I1204 10:36:45.210135 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"20dd5c87-1885-44c6-8ed4-97898e603817","Type":"ContainerStarted","Data":"052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d"} Dec 04 10:36:45 crc kubenswrapper[4943]: I1204 10:36:45.210717 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"20dd5c87-1885-44c6-8ed4-97898e603817","Type":"ContainerStarted","Data":"07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792"} Dec 04 10:36:45 crc kubenswrapper[4943]: I1204 10:36:45.210729 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"20dd5c87-1885-44c6-8ed4-97898e603817","Type":"ContainerStarted","Data":"9c34272e95096ed3f87fb07aa12984a657b7adfee180fcfde87417cf31709098"} Dec 04 10:36:45 crc kubenswrapper[4943]: I1204 10:36:45.244424 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-2" podStartSLOduration=1.244398275 podStartE2EDuration="1.244398275s" podCreationTimestamp="2025-12-04 10:36:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:36:45.239783697 +0000 UTC m=+1253.828759565" watchObservedRunningTime="2025-12-04 10:36:45.244398275 +0000 UTC m=+1253.833374143" Dec 04 10:36:48 crc kubenswrapper[4943]: I1204 10:36:48.564635 4943 scope.go:117] "RemoveContainer" containerID="773d56d8459293541800c4ad4e06c3d84b13851e31fdf7aaa2de8e84e38749e8" Dec 04 10:36:48 crc kubenswrapper[4943]: I1204 10:36:48.564968 4943 scope.go:117] "RemoveContainer" containerID="9dbe4b564c22dcdc28c5907566e1deee1d2273994bc376456b1a2f567a7d4255" Dec 04 10:36:49 crc kubenswrapper[4943]: I1204 10:36:49.239573 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerStarted","Data":"a5e05e77639a143c450e901da7187ae9ef9fe11a14e7b0d28d9484b7f49c0a68"} Dec 04 10:36:49 crc kubenswrapper[4943]: I1204 10:36:49.240257 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerStarted","Data":"cdb78185274d9ca800f5449ab5d23529653eb05dfc50d6507d84c36a3415a769"} Dec 04 10:36:49 crc kubenswrapper[4943]: I1204 10:36:49.422419 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:49 crc kubenswrapper[4943]: I1204 10:36:49.622733 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:50 crc kubenswrapper[4943]: I1204 10:36:50.363003 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:36:51 crc kubenswrapper[4943]: I1204 10:36:51.043656 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 04 10:36:51 crc kubenswrapper[4943]: I1204 10:36:51.256225 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerDied","Data":"a5e05e77639a143c450e901da7187ae9ef9fe11a14e7b0d28d9484b7f49c0a68"} Dec 04 10:36:51 crc kubenswrapper[4943]: I1204 10:36:51.256229 4943 generic.go:334] "Generic (PLEG): container finished" podID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerID="a5e05e77639a143c450e901da7187ae9ef9fe11a14e7b0d28d9484b7f49c0a68" exitCode=1 Dec 04 10:36:51 crc kubenswrapper[4943]: I1204 10:36:51.256292 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerDied","Data":"cdb78185274d9ca800f5449ab5d23529653eb05dfc50d6507d84c36a3415a769"} Dec 04 10:36:51 crc kubenswrapper[4943]: I1204 10:36:51.256312 4943 generic.go:334] "Generic (PLEG): container finished" podID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerID="cdb78185274d9ca800f5449ab5d23529653eb05dfc50d6507d84c36a3415a769" exitCode=1 Dec 04 10:36:51 crc kubenswrapper[4943]: I1204 10:36:51.256360 4943 scope.go:117] "RemoveContainer" containerID="9dbe4b564c22dcdc28c5907566e1deee1d2273994bc376456b1a2f567a7d4255" Dec 04 10:36:51 crc kubenswrapper[4943]: I1204 10:36:51.256894 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-2" podUID="20dd5c87-1885-44c6-8ed4-97898e603817" containerName="cinder-backup" containerID="cri-o://07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792" gracePeriod=30 Dec 04 10:36:51 crc kubenswrapper[4943]: I1204 10:36:51.257070 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-2" podUID="20dd5c87-1885-44c6-8ed4-97898e603817" containerName="probe" containerID="cri-o://052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d" gracePeriod=30 Dec 04 10:36:51 crc kubenswrapper[4943]: I1204 10:36:51.257121 4943 scope.go:117] "RemoveContainer" containerID="cdb78185274d9ca800f5449ab5d23529653eb05dfc50d6507d84c36a3415a769" Dec 04 10:36:51 crc kubenswrapper[4943]: I1204 10:36:51.257230 4943 scope.go:117] "RemoveContainer" containerID="a5e05e77639a143c450e901da7187ae9ef9fe11a14e7b0d28d9484b7f49c0a68" Dec 04 10:36:51 crc kubenswrapper[4943]: E1204 10:36:51.257561 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:36:51 crc kubenswrapper[4943]: I1204 10:36:51.304960 4943 scope.go:117] "RemoveContainer" containerID="773d56d8459293541800c4ad4e06c3d84b13851e31fdf7aaa2de8e84e38749e8" Dec 04 10:36:52 crc kubenswrapper[4943]: I1204 10:36:52.267327 4943 generic.go:334] "Generic (PLEG): container finished" podID="20dd5c87-1885-44c6-8ed4-97898e603817" containerID="052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d" exitCode=0 Dec 04 10:36:52 crc kubenswrapper[4943]: I1204 10:36:52.267400 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"20dd5c87-1885-44c6-8ed4-97898e603817","Type":"ContainerDied","Data":"052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d"} Dec 04 10:36:52 crc kubenswrapper[4943]: I1204 10:36:52.270871 4943 scope.go:117] "RemoveContainer" containerID="cdb78185274d9ca800f5449ab5d23529653eb05dfc50d6507d84c36a3415a769" Dec 04 10:36:52 crc kubenswrapper[4943]: I1204 10:36:52.270968 4943 scope.go:117] "RemoveContainer" containerID="a5e05e77639a143c450e901da7187ae9ef9fe11a14e7b0d28d9484b7f49c0a68" Dec 04 10:36:52 crc kubenswrapper[4943]: E1204 10:36:52.271263 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:36:53 crc kubenswrapper[4943]: I1204 10:36:53.363482 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:36:53 crc kubenswrapper[4943]: I1204 10:36:53.365131 4943 scope.go:117] "RemoveContainer" containerID="cdb78185274d9ca800f5449ab5d23529653eb05dfc50d6507d84c36a3415a769" Dec 04 10:36:53 crc kubenswrapper[4943]: I1204 10:36:53.365275 4943 scope.go:117] "RemoveContainer" containerID="a5e05e77639a143c450e901da7187ae9ef9fe11a14e7b0d28d9484b7f49c0a68" Dec 04 10:36:53 crc kubenswrapper[4943]: E1204 10:36:53.365708 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.362477 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.363394 4943 scope.go:117] "RemoveContainer" containerID="cdb78185274d9ca800f5449ab5d23529653eb05dfc50d6507d84c36a3415a769" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.363413 4943 scope.go:117] "RemoveContainer" containerID="a5e05e77639a143c450e901da7187ae9ef9fe11a14e7b0d28d9484b7f49c0a68" Dec 04 10:36:55 crc kubenswrapper[4943]: E1204 10:36:55.363778 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.637711 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731149 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rl6bc\" (UniqueName: \"kubernetes.io/projected/20dd5c87-1885-44c6-8ed4-97898e603817-kube-api-access-rl6bc\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731265 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-run\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731289 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-lib-modules\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731337 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-config-data\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731356 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-locks-cinder\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731374 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-dev\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731387 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-machine-id\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731403 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-sys\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731396 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-run" (OuterVolumeSpecName: "run") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731456 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-scripts\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731464 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731498 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-dev" (OuterVolumeSpecName: "dev") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731486 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-nvme\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731525 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731546 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731568 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731552 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-sys" (OuterVolumeSpecName: "sys") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731610 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-config-data-custom\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731640 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-locks-brick\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731661 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-lib-cinder\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.731712 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-iscsi\") pod \"20dd5c87-1885-44c6-8ed4-97898e603817\" (UID: \"20dd5c87-1885-44c6-8ed4-97898e603817\") " Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.732050 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.732094 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.732281 4943 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.732307 4943 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-run\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.732316 4943 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.732325 4943 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.732335 4943 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-dev\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.732343 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.732351 4943 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-sys\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.732359 4943 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.732368 4943 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.732277 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.737522 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.737809 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20dd5c87-1885-44c6-8ed4-97898e603817-kube-api-access-rl6bc" (OuterVolumeSpecName: "kube-api-access-rl6bc") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "kube-api-access-rl6bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.741357 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-scripts" (OuterVolumeSpecName: "scripts") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.796271 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-config-data" (OuterVolumeSpecName: "config-data") pod "20dd5c87-1885-44c6-8ed4-97898e603817" (UID: "20dd5c87-1885-44c6-8ed4-97898e603817"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.833001 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rl6bc\" (UniqueName: \"kubernetes.io/projected/20dd5c87-1885-44c6-8ed4-97898e603817-kube-api-access-rl6bc\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.833043 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.833074 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.833087 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20dd5c87-1885-44c6-8ed4-97898e603817-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:55 crc kubenswrapper[4943]: I1204 10:36:55.833097 4943 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/20dd5c87-1885-44c6-8ed4-97898e603817-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.295418 4943 generic.go:334] "Generic (PLEG): container finished" podID="20dd5c87-1885-44c6-8ed4-97898e603817" containerID="07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792" exitCode=0 Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.295523 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-2" Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.295529 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"20dd5c87-1885-44c6-8ed4-97898e603817","Type":"ContainerDied","Data":"07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792"} Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.295855 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-2" event={"ID":"20dd5c87-1885-44c6-8ed4-97898e603817","Type":"ContainerDied","Data":"9c34272e95096ed3f87fb07aa12984a657b7adfee180fcfde87417cf31709098"} Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.295876 4943 scope.go:117] "RemoveContainer" containerID="052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d" Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.324980 4943 scope.go:117] "RemoveContainer" containerID="07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792" Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.333794 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.339963 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-2"] Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.345288 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.345582 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-1" podUID="f44de5e7-3753-4159-b01c-970889c7adf7" containerName="cinder-backup" containerID="cri-o://f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969" gracePeriod=30 Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.345685 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-1" podUID="f44de5e7-3753-4159-b01c-970889c7adf7" containerName="probe" containerID="cri-o://95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441" gracePeriod=30 Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.352438 4943 scope.go:117] "RemoveContainer" containerID="052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d" Dec 04 10:36:56 crc kubenswrapper[4943]: E1204 10:36:56.353001 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d\": container with ID starting with 052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d not found: ID does not exist" containerID="052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d" Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.353083 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d"} err="failed to get container status \"052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d\": rpc error: code = NotFound desc = could not find container \"052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d\": container with ID starting with 052ff5bf442a42053a2a8ca4e43f12e85286f83c00e06412b11619a5ed93c96d not found: ID does not exist" Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.353133 4943 scope.go:117] "RemoveContainer" containerID="07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792" Dec 04 10:36:56 crc kubenswrapper[4943]: E1204 10:36:56.353558 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792\": container with ID starting with 07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792 not found: ID does not exist" containerID="07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792" Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.353659 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792"} err="failed to get container status \"07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792\": rpc error: code = NotFound desc = could not find container \"07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792\": container with ID starting with 07060688702caf1a572e307e1e080ab365ee76c0b06d314c079f4a30372e6792 not found: ID does not exist" Dec 04 10:36:56 crc kubenswrapper[4943]: I1204 10:36:56.572528 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20dd5c87-1885-44c6-8ed4-97898e603817" path="/var/lib/kubelet/pods/20dd5c87-1885-44c6-8ed4-97898e603817/volumes" Dec 04 10:36:57 crc kubenswrapper[4943]: I1204 10:36:57.307501 4943 generic.go:334] "Generic (PLEG): container finished" podID="f44de5e7-3753-4159-b01c-970889c7adf7" containerID="95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441" exitCode=0 Dec 04 10:36:57 crc kubenswrapper[4943]: I1204 10:36:57.307552 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"f44de5e7-3753-4159-b01c-970889c7adf7","Type":"ContainerDied","Data":"95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441"} Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.300090 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.344048 4943 generic.go:334] "Generic (PLEG): container finished" podID="f44de5e7-3753-4159-b01c-970889c7adf7" containerID="f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969" exitCode=0 Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.344110 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"f44de5e7-3753-4159-b01c-970889c7adf7","Type":"ContainerDied","Data":"f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969"} Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.344143 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-1" event={"ID":"f44de5e7-3753-4159-b01c-970889c7adf7","Type":"ContainerDied","Data":"c41bb7a39cadc1137089c6f0c38c3d1cffbf0c5549bc4bfdc8aeab37badc99ae"} Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.344166 4943 scope.go:117] "RemoveContainer" containerID="95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.344111 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-1" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.366676 4943 scope.go:117] "RemoveContainer" containerID="f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.387590 4943 scope.go:117] "RemoveContainer" containerID="95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441" Dec 04 10:37:01 crc kubenswrapper[4943]: E1204 10:37:01.387998 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441\": container with ID starting with 95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441 not found: ID does not exist" containerID="95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.388040 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441"} err="failed to get container status \"95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441\": rpc error: code = NotFound desc = could not find container \"95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441\": container with ID starting with 95ea3aa49c828fca0978cc0740a1c54dc29d3cb54cddf4d90167a9e11d36c441 not found: ID does not exist" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.388084 4943 scope.go:117] "RemoveContainer" containerID="f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969" Dec 04 10:37:01 crc kubenswrapper[4943]: E1204 10:37:01.388796 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969\": container with ID starting with f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969 not found: ID does not exist" containerID="f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.388817 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969"} err="failed to get container status \"f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969\": rpc error: code = NotFound desc = could not find container \"f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969\": container with ID starting with f88b34d63811d8f3f57fe20e9ac2fe2e7998fbb2e7749ce551f029ed97de5969 not found: ID does not exist" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.423416 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-lib-cinder\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.423481 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-config-data\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.423524 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-scripts\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.423552 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-config-data-custom\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.423564 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.423578 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2265w\" (UniqueName: \"kubernetes.io/projected/f44de5e7-3753-4159-b01c-970889c7adf7-kube-api-access-2265w\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.423763 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-run\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.423840 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-run" (OuterVolumeSpecName: "run") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.423893 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-sys\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.423970 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-sys" (OuterVolumeSpecName: "sys") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424005 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-locks-cinder\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424034 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-lib-modules\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424136 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-locks-brick\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424186 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-iscsi\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424235 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-nvme\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424263 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-dev\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424297 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-machine-id\") pod \"f44de5e7-3753-4159-b01c-970889c7adf7\" (UID: \"f44de5e7-3753-4159-b01c-970889c7adf7\") " Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424069 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424396 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424407 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424082 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424448 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-dev" (OuterVolumeSpecName: "dev") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424493 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.424532 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.425010 4943 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.425057 4943 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-run\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.425076 4943 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-sys\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.425088 4943 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.425176 4943 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.425190 4943 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.425217 4943 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.425229 4943 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.425240 4943 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-dev\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.425250 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f44de5e7-3753-4159-b01c-970889c7adf7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.428808 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.429876 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-scripts" (OuterVolumeSpecName: "scripts") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.429895 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f44de5e7-3753-4159-b01c-970889c7adf7-kube-api-access-2265w" (OuterVolumeSpecName: "kube-api-access-2265w") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "kube-api-access-2265w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.513994 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-config-data" (OuterVolumeSpecName: "config-data") pod "f44de5e7-3753-4159-b01c-970889c7adf7" (UID: "f44de5e7-3753-4159-b01c-970889c7adf7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.527098 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.527128 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.527137 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f44de5e7-3753-4159-b01c-970889c7adf7-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.527147 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2265w\" (UniqueName: \"kubernetes.io/projected/f44de5e7-3753-4159-b01c-970889c7adf7-kube-api-access-2265w\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.677603 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 04 10:37:01 crc kubenswrapper[4943]: I1204 10:37:01.682614 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-1"] Dec 04 10:37:02 crc kubenswrapper[4943]: I1204 10:37:02.411303 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:02 crc kubenswrapper[4943]: I1204 10:37:02.411660 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" containerName="cinder-api-log" containerID="cri-o://7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf" gracePeriod=30 Dec 04 10:37:02 crc kubenswrapper[4943]: I1204 10:37:02.411697 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" containerName="cinder-api" containerID="cri-o://32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1" gracePeriod=30 Dec 04 10:37:02 crc kubenswrapper[4943]: I1204 10:37:02.596388 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f44de5e7-3753-4159-b01c-970889c7adf7" path="/var/lib/kubelet/pods/f44de5e7-3753-4159-b01c-970889c7adf7/volumes" Dec 04 10:37:03 crc kubenswrapper[4943]: I1204 10:37:03.361344 4943 generic.go:334] "Generic (PLEG): container finished" podID="f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" containerID="7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf" exitCode=143 Dec 04 10:37:03 crc kubenswrapper[4943]: I1204 10:37:03.361428 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e","Type":"ContainerDied","Data":"7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf"} Dec 04 10:37:05 crc kubenswrapper[4943]: I1204 10:37:05.516764 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-0" podUID="f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.91:8776/healthcheck\": read tcp 10.217.0.2:44262->10.217.0.91:8776: read: connection reset by peer" Dec 04 10:37:05 crc kubenswrapper[4943]: I1204 10:37:05.940229 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.052679 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-etc-machine-id\") pod \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.052773 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-config-data\") pod \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.052794 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-config-data-custom\") pod \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.052830 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" (UID: "f8489a8d-2b17-4a37-b6f8-0da9b6a2318e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.052893 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdrbm\" (UniqueName: \"kubernetes.io/projected/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-kube-api-access-fdrbm\") pod \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.052942 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-scripts\") pod \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.052963 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-logs\") pod \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\" (UID: \"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e\") " Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.053262 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.053808 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-logs" (OuterVolumeSpecName: "logs") pod "f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" (UID: "f8489a8d-2b17-4a37-b6f8-0da9b6a2318e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.058313 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-scripts" (OuterVolumeSpecName: "scripts") pod "f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" (UID: "f8489a8d-2b17-4a37-b6f8-0da9b6a2318e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.063610 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-kube-api-access-fdrbm" (OuterVolumeSpecName: "kube-api-access-fdrbm") pod "f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" (UID: "f8489a8d-2b17-4a37-b6f8-0da9b6a2318e"). InnerVolumeSpecName "kube-api-access-fdrbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.075713 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" (UID: "f8489a8d-2b17-4a37-b6f8-0da9b6a2318e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.096450 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-config-data" (OuterVolumeSpecName: "config-data") pod "f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" (UID: "f8489a8d-2b17-4a37-b6f8-0da9b6a2318e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.155257 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.155295 4943 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-logs\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.155304 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.155313 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.155324 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdrbm\" (UniqueName: \"kubernetes.io/projected/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e-kube-api-access-fdrbm\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.384754 4943 generic.go:334] "Generic (PLEG): container finished" podID="f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" containerID="32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1" exitCode=0 Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.384817 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e","Type":"ContainerDied","Data":"32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1"} Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.384852 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"f8489a8d-2b17-4a37-b6f8-0da9b6a2318e","Type":"ContainerDied","Data":"0a0a8f3e4c81bf96c99200a2dea237d6c5bf65bd62fa294b56be18a779500705"} Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.384849 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.384873 4943 scope.go:117] "RemoveContainer" containerID="32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.411304 4943 scope.go:117] "RemoveContainer" containerID="7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.433435 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.440685 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.441861 4943 scope.go:117] "RemoveContainer" containerID="32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1" Dec 04 10:37:06 crc kubenswrapper[4943]: E1204 10:37:06.442385 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1\": container with ID starting with 32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1 not found: ID does not exist" containerID="32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.442418 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1"} err="failed to get container status \"32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1\": rpc error: code = NotFound desc = could not find container \"32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1\": container with ID starting with 32427fb79c3757f9e81a5492c09f9b50b49444297c23575f1c339a31ba05fcd1 not found: ID does not exist" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.442446 4943 scope.go:117] "RemoveContainer" containerID="7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf" Dec 04 10:37:06 crc kubenswrapper[4943]: E1204 10:37:06.442980 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf\": container with ID starting with 7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf not found: ID does not exist" containerID="7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.443024 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf"} err="failed to get container status \"7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf\": rpc error: code = NotFound desc = could not find container \"7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf\": container with ID starting with 7b27ec2d17649429c15086bda9437740ae024f7055f2a3fa3f9459476103ebaf not found: ID does not exist" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.572892 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" path="/var/lib/kubelet/pods/f8489a8d-2b17-4a37-b6f8-0da9b6a2318e/volumes" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.645213 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:06 crc kubenswrapper[4943]: E1204 10:37:06.645792 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" containerName="cinder-api" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.645822 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" containerName="cinder-api" Dec 04 10:37:06 crc kubenswrapper[4943]: E1204 10:37:06.645831 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20dd5c87-1885-44c6-8ed4-97898e603817" containerName="cinder-backup" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.645837 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="20dd5c87-1885-44c6-8ed4-97898e603817" containerName="cinder-backup" Dec 04 10:37:06 crc kubenswrapper[4943]: E1204 10:37:06.645882 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f44de5e7-3753-4159-b01c-970889c7adf7" containerName="cinder-backup" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.645892 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f44de5e7-3753-4159-b01c-970889c7adf7" containerName="cinder-backup" Dec 04 10:37:06 crc kubenswrapper[4943]: E1204 10:37:06.645907 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f44de5e7-3753-4159-b01c-970889c7adf7" containerName="probe" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.645913 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f44de5e7-3753-4159-b01c-970889c7adf7" containerName="probe" Dec 04 10:37:06 crc kubenswrapper[4943]: E1204 10:37:06.645919 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20dd5c87-1885-44c6-8ed4-97898e603817" containerName="probe" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.645925 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="20dd5c87-1885-44c6-8ed4-97898e603817" containerName="probe" Dec 04 10:37:06 crc kubenswrapper[4943]: E1204 10:37:06.645934 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" containerName="cinder-api-log" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.645940 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" containerName="cinder-api-log" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.646088 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="20dd5c87-1885-44c6-8ed4-97898e603817" containerName="cinder-backup" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.646101 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="20dd5c87-1885-44c6-8ed4-97898e603817" containerName="probe" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.646112 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f44de5e7-3753-4159-b01c-970889c7adf7" containerName="probe" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.646120 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" containerName="cinder-api" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.646127 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8489a8d-2b17-4a37-b6f8-0da9b6a2318e" containerName="cinder-api-log" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.646140 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f44de5e7-3753-4159-b01c-970889c7adf7" containerName="cinder-backup" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.646888 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.650452 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.657700 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.666504 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.667602 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.676134 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.677732 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.691375 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.695819 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.763568 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99251963-b5ca-4bc8-a37b-e38d8f2f72db-logs\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.763623 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99251963-b5ca-4bc8-a37b-e38d8f2f72db-etc-machine-id\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.763662 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-config-data\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.763679 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-scripts\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.763703 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv7wg\" (UniqueName: \"kubernetes.io/projected/99251963-b5ca-4bc8-a37b-e38d8f2f72db-kube-api-access-bv7wg\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.764401 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-config-data-custom\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.866174 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-config-data-custom\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.866445 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-config-data\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.866476 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-config-data\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867015 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-config-data-custom\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867078 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c54f766-d384-46e5-a29f-2d1ac4181ba6-logs\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867110 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99251963-b5ca-4bc8-a37b-e38d8f2f72db-logs\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867127 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99251963-b5ca-4bc8-a37b-e38d8f2f72db-etc-machine-id\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867154 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-config-data\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867169 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-scripts\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867185 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-config-data-custom\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867446 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99251963-b5ca-4bc8-a37b-e38d8f2f72db-etc-machine-id\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867220 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv7wg\" (UniqueName: \"kubernetes.io/projected/99251963-b5ca-4bc8-a37b-e38d8f2f72db-kube-api-access-bv7wg\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867534 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-scripts\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867551 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-scripts\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867593 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76jt5\" (UniqueName: \"kubernetes.io/projected/9de47e9d-5a4c-4563-b419-5ee9ca102f08-kube-api-access-76jt5\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867651 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9de47e9d-5a4c-4563-b419-5ee9ca102f08-etc-machine-id\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867666 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9de47e9d-5a4c-4563-b419-5ee9ca102f08-logs\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867679 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c54f766-d384-46e5-a29f-2d1ac4181ba6-etc-machine-id\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867759 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jdvj\" (UniqueName: \"kubernetes.io/projected/0c54f766-d384-46e5-a29f-2d1ac4181ba6-kube-api-access-4jdvj\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.867861 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99251963-b5ca-4bc8-a37b-e38d8f2f72db-logs\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.872016 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-config-data\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.873248 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-config-data-custom\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.884524 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-scripts\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.885239 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv7wg\" (UniqueName: \"kubernetes.io/projected/99251963-b5ca-4bc8-a37b-e38d8f2f72db-kube-api-access-bv7wg\") pod \"cinder-api-0\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.969414 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-config-data-custom\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.969518 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-scripts\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.969551 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-scripts\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.969605 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76jt5\" (UniqueName: \"kubernetes.io/projected/9de47e9d-5a4c-4563-b419-5ee9ca102f08-kube-api-access-76jt5\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.969648 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9de47e9d-5a4c-4563-b419-5ee9ca102f08-etc-machine-id\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.969682 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9de47e9d-5a4c-4563-b419-5ee9ca102f08-logs\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.969721 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c54f766-d384-46e5-a29f-2d1ac4181ba6-etc-machine-id\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.969872 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c54f766-d384-46e5-a29f-2d1ac4181ba6-etc-machine-id\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.969810 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9de47e9d-5a4c-4563-b419-5ee9ca102f08-etc-machine-id\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.970524 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9de47e9d-5a4c-4563-b419-5ee9ca102f08-logs\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.970869 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jdvj\" (UniqueName: \"kubernetes.io/projected/0c54f766-d384-46e5-a29f-2d1ac4181ba6-kube-api-access-4jdvj\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.971565 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-config-data\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.971644 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-config-data\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.971717 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-config-data-custom\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.972529 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c54f766-d384-46e5-a29f-2d1ac4181ba6-logs\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.973188 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-scripts\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.973272 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c54f766-d384-46e5-a29f-2d1ac4181ba6-logs\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.976629 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-config-data-custom\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.977264 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-config-data\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.978042 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-config-data-custom\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.980453 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-scripts\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.980536 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-config-data\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.989985 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76jt5\" (UniqueName: \"kubernetes.io/projected/9de47e9d-5a4c-4563-b419-5ee9ca102f08-kube-api-access-76jt5\") pod \"cinder-api-1\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:06 crc kubenswrapper[4943]: I1204 10:37:06.990550 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:07 crc kubenswrapper[4943]: I1204 10:37:07.004867 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jdvj\" (UniqueName: \"kubernetes.io/projected/0c54f766-d384-46e5-a29f-2d1ac4181ba6-kube-api-access-4jdvj\") pod \"cinder-api-2\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:07 crc kubenswrapper[4943]: I1204 10:37:07.017257 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:07 crc kubenswrapper[4943]: I1204 10:37:07.248181 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 04 10:37:07 crc kubenswrapper[4943]: I1204 10:37:07.302129 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:07 crc kubenswrapper[4943]: I1204 10:37:07.396117 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:07 crc kubenswrapper[4943]: I1204 10:37:07.411398 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"9de47e9d-5a4c-4563-b419-5ee9ca102f08","Type":"ContainerStarted","Data":"2eebbd864927ecd09c680823d4d017b0e1118cfcb17e687dc26ed8babaa07459"} Dec 04 10:37:07 crc kubenswrapper[4943]: I1204 10:37:07.565334 4943 scope.go:117] "RemoveContainer" containerID="cdb78185274d9ca800f5449ab5d23529653eb05dfc50d6507d84c36a3415a769" Dec 04 10:37:07 crc kubenswrapper[4943]: I1204 10:37:07.565366 4943 scope.go:117] "RemoveContainer" containerID="a5e05e77639a143c450e901da7187ae9ef9fe11a14e7b0d28d9484b7f49c0a68" Dec 04 10:37:07 crc kubenswrapper[4943]: E1204 10:37:07.565624 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:37:07 crc kubenswrapper[4943]: I1204 10:37:07.733592 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 04 10:37:07 crc kubenswrapper[4943]: W1204 10:37:07.741849 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c54f766_d384_46e5_a29f_2d1ac4181ba6.slice/crio-b734712b8f614a57185333054b5bc0b1d67b6be062141e9dced6db7b1e94a5bc WatchSource:0}: Error finding container b734712b8f614a57185333054b5bc0b1d67b6be062141e9dced6db7b1e94a5bc: Status 404 returned error can't find the container with id b734712b8f614a57185333054b5bc0b1d67b6be062141e9dced6db7b1e94a5bc Dec 04 10:37:08 crc kubenswrapper[4943]: I1204 10:37:08.436595 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"9de47e9d-5a4c-4563-b419-5ee9ca102f08","Type":"ContainerStarted","Data":"a5ecee7ed59649a14a5215ade747df8815db39dc12ef3a2ad5bc0fe6e1475653"} Dec 04 10:37:08 crc kubenswrapper[4943]: I1204 10:37:08.437245 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:08 crc kubenswrapper[4943]: I1204 10:37:08.437266 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"9de47e9d-5a4c-4563-b419-5ee9ca102f08","Type":"ContainerStarted","Data":"4743f6093002a405c3c27f10ca16dd54e64b2b362666f710735fb3fc52c59e7a"} Dec 04 10:37:08 crc kubenswrapper[4943]: I1204 10:37:08.440114 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"0c54f766-d384-46e5-a29f-2d1ac4181ba6","Type":"ContainerStarted","Data":"a6ad19bb3ab132a00ade526b019e2fd89bffbfc25b38e8a321bec0af62d93f00"} Dec 04 10:37:08 crc kubenswrapper[4943]: I1204 10:37:08.440160 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"0c54f766-d384-46e5-a29f-2d1ac4181ba6","Type":"ContainerStarted","Data":"b734712b8f614a57185333054b5bc0b1d67b6be062141e9dced6db7b1e94a5bc"} Dec 04 10:37:08 crc kubenswrapper[4943]: I1204 10:37:08.443168 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"99251963-b5ca-4bc8-a37b-e38d8f2f72db","Type":"ContainerStarted","Data":"cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb"} Dec 04 10:37:08 crc kubenswrapper[4943]: I1204 10:37:08.443238 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"99251963-b5ca-4bc8-a37b-e38d8f2f72db","Type":"ContainerStarted","Data":"9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5"} Dec 04 10:37:08 crc kubenswrapper[4943]: I1204 10:37:08.443251 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"99251963-b5ca-4bc8-a37b-e38d8f2f72db","Type":"ContainerStarted","Data":"fe5d5c271c5499d9e3d87ff1aa172c1ca1deea9242ca3384fff35337092152ce"} Dec 04 10:37:08 crc kubenswrapper[4943]: I1204 10:37:08.443435 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:08 crc kubenswrapper[4943]: I1204 10:37:08.458428 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-1" podStartSLOduration=2.4581862660000002 podStartE2EDuration="2.458186266s" podCreationTimestamp="2025-12-04 10:37:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:37:08.451499992 +0000 UTC m=+1277.040475860" watchObservedRunningTime="2025-12-04 10:37:08.458186266 +0000 UTC m=+1277.047162134" Dec 04 10:37:08 crc kubenswrapper[4943]: I1204 10:37:08.480022 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=2.480000786 podStartE2EDuration="2.480000786s" podCreationTimestamp="2025-12-04 10:37:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:37:08.470772923 +0000 UTC m=+1277.059748791" watchObservedRunningTime="2025-12-04 10:37:08.480000786 +0000 UTC m=+1277.068976654" Dec 04 10:37:09 crc kubenswrapper[4943]: I1204 10:37:09.451776 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"0c54f766-d384-46e5-a29f-2d1ac4181ba6","Type":"ContainerStarted","Data":"1df6307739aed566ad0c5b08aaa675bc773c49be4f9c6c4e519561a97e23defb"} Dec 04 10:37:09 crc kubenswrapper[4943]: I1204 10:37:09.453027 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:09 crc kubenswrapper[4943]: I1204 10:37:09.481760 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-2" podStartSLOduration=3.481736757 podStartE2EDuration="3.481736757s" podCreationTimestamp="2025-12-04 10:37:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:37:09.468553504 +0000 UTC m=+1278.057529402" watchObservedRunningTime="2025-12-04 10:37:09.481736757 +0000 UTC m=+1278.070712635" Dec 04 10:37:18 crc kubenswrapper[4943]: I1204 10:37:18.832408 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:18 crc kubenswrapper[4943]: I1204 10:37:18.848365 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:19 crc kubenswrapper[4943]: I1204 10:37:19.317874 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.015806 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.016453 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-2" podUID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" containerName="cinder-api-log" containerID="cri-o://a6ad19bb3ab132a00ade526b019e2fd89bffbfc25b38e8a321bec0af62d93f00" gracePeriod=30 Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.016932 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-2" podUID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" containerName="cinder-api" containerID="cri-o://1df6307739aed566ad0c5b08aaa675bc773c49be4f9c6c4e519561a97e23defb" gracePeriod=30 Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.026089 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="cinder-kuttl-tests/cinder-api-2" podUID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.98:8776/healthcheck\": EOF" Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.026283 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.026529 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-1" podUID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" containerName="cinder-api-log" containerID="cri-o://4743f6093002a405c3c27f10ca16dd54e64b2b362666f710735fb3fc52c59e7a" gracePeriod=30 Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.026744 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-1" podUID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" containerName="cinder-api" containerID="cri-o://a5ecee7ed59649a14a5215ade747df8815db39dc12ef3a2ad5bc0fe6e1475653" gracePeriod=30 Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.043024 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="cinder-kuttl-tests/cinder-api-1" podUID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.99:8776/healthcheck\": EOF" Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.543296 4943 generic.go:334] "Generic (PLEG): container finished" podID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" containerID="4743f6093002a405c3c27f10ca16dd54e64b2b362666f710735fb3fc52c59e7a" exitCode=143 Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.543373 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"9de47e9d-5a4c-4563-b419-5ee9ca102f08","Type":"ContainerDied","Data":"4743f6093002a405c3c27f10ca16dd54e64b2b362666f710735fb3fc52c59e7a"} Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.546950 4943 generic.go:334] "Generic (PLEG): container finished" podID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" containerID="a6ad19bb3ab132a00ade526b019e2fd89bffbfc25b38e8a321bec0af62d93f00" exitCode=143 Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.547019 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"0c54f766-d384-46e5-a29f-2d1ac4181ba6","Type":"ContainerDied","Data":"a6ad19bb3ab132a00ade526b019e2fd89bffbfc25b38e8a321bec0af62d93f00"} Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.565376 4943 scope.go:117] "RemoveContainer" containerID="cdb78185274d9ca800f5449ab5d23529653eb05dfc50d6507d84c36a3415a769" Dec 04 10:37:20 crc kubenswrapper[4943]: I1204 10:37:20.565417 4943 scope.go:117] "RemoveContainer" containerID="a5e05e77639a143c450e901da7187ae9ef9fe11a14e7b0d28d9484b7f49c0a68" Dec 04 10:37:20 crc kubenswrapper[4943]: E1204 10:37:20.565844 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(5dbc6cc7-a807-4f74-ad16-feecbab07846)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" Dec 04 10:37:24 crc kubenswrapper[4943]: I1204 10:37:24.440195 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-2" podUID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.98:8776/healthcheck\": read tcp 10.217.0.2:55266->10.217.0.98:8776: read: connection reset by peer" Dec 04 10:37:24 crc kubenswrapper[4943]: I1204 10:37:24.441508 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-1" podUID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.99:8776/healthcheck\": read tcp 10.217.0.2:58024->10.217.0.99:8776: read: connection reset by peer" Dec 04 10:37:24 crc kubenswrapper[4943]: I1204 10:37:24.597799 4943 generic.go:334] "Generic (PLEG): container finished" podID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" containerID="a5ecee7ed59649a14a5215ade747df8815db39dc12ef3a2ad5bc0fe6e1475653" exitCode=0 Dec 04 10:37:24 crc kubenswrapper[4943]: I1204 10:37:24.597890 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"9de47e9d-5a4c-4563-b419-5ee9ca102f08","Type":"ContainerDied","Data":"a5ecee7ed59649a14a5215ade747df8815db39dc12ef3a2ad5bc0fe6e1475653"} Dec 04 10:37:24 crc kubenswrapper[4943]: I1204 10:37:24.604938 4943 generic.go:334] "Generic (PLEG): container finished" podID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" containerID="1df6307739aed566ad0c5b08aaa675bc773c49be4f9c6c4e519561a97e23defb" exitCode=0 Dec 04 10:37:24 crc kubenswrapper[4943]: I1204 10:37:24.604991 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"0c54f766-d384-46e5-a29f-2d1ac4181ba6","Type":"ContainerDied","Data":"1df6307739aed566ad0c5b08aaa675bc773c49be4f9c6c4e519561a97e23defb"} Dec 04 10:37:24 crc kubenswrapper[4943]: I1204 10:37:24.887210 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:24 crc kubenswrapper[4943]: I1204 10:37:24.896604 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.074687 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-config-data-custom\") pod \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.074767 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jdvj\" (UniqueName: \"kubernetes.io/projected/0c54f766-d384-46e5-a29f-2d1ac4181ba6-kube-api-access-4jdvj\") pod \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.074827 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-scripts\") pod \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.074890 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-config-data\") pod \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.074909 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9de47e9d-5a4c-4563-b419-5ee9ca102f08-etc-machine-id\") pod \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.074943 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-config-data\") pod \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.074973 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c54f766-d384-46e5-a29f-2d1ac4181ba6-logs\") pod \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.075033 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c54f766-d384-46e5-a29f-2d1ac4181ba6-etc-machine-id\") pod \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.075066 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-config-data-custom\") pod \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.075086 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-scripts\") pod \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\" (UID: \"0c54f766-d384-46e5-a29f-2d1ac4181ba6\") " Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.075119 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9de47e9d-5a4c-4563-b419-5ee9ca102f08-logs\") pod \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.075141 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76jt5\" (UniqueName: \"kubernetes.io/projected/9de47e9d-5a4c-4563-b419-5ee9ca102f08-kube-api-access-76jt5\") pod \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\" (UID: \"9de47e9d-5a4c-4563-b419-5ee9ca102f08\") " Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.075160 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0c54f766-d384-46e5-a29f-2d1ac4181ba6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0c54f766-d384-46e5-a29f-2d1ac4181ba6" (UID: "0c54f766-d384-46e5-a29f-2d1ac4181ba6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.075175 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9de47e9d-5a4c-4563-b419-5ee9ca102f08-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9de47e9d-5a4c-4563-b419-5ee9ca102f08" (UID: "9de47e9d-5a4c-4563-b419-5ee9ca102f08"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.075500 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9de47e9d-5a4c-4563-b419-5ee9ca102f08-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.075517 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c54f766-d384-46e5-a29f-2d1ac4181ba6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.075625 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9de47e9d-5a4c-4563-b419-5ee9ca102f08-logs" (OuterVolumeSpecName: "logs") pod "9de47e9d-5a4c-4563-b419-5ee9ca102f08" (UID: "9de47e9d-5a4c-4563-b419-5ee9ca102f08"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.075706 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c54f766-d384-46e5-a29f-2d1ac4181ba6-logs" (OuterVolumeSpecName: "logs") pod "0c54f766-d384-46e5-a29f-2d1ac4181ba6" (UID: "0c54f766-d384-46e5-a29f-2d1ac4181ba6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.081495 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0c54f766-d384-46e5-a29f-2d1ac4181ba6" (UID: "0c54f766-d384-46e5-a29f-2d1ac4181ba6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.081521 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9de47e9d-5a4c-4563-b419-5ee9ca102f08-kube-api-access-76jt5" (OuterVolumeSpecName: "kube-api-access-76jt5") pod "9de47e9d-5a4c-4563-b419-5ee9ca102f08" (UID: "9de47e9d-5a4c-4563-b419-5ee9ca102f08"). InnerVolumeSpecName "kube-api-access-76jt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.081602 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-scripts" (OuterVolumeSpecName: "scripts") pod "9de47e9d-5a4c-4563-b419-5ee9ca102f08" (UID: "9de47e9d-5a4c-4563-b419-5ee9ca102f08"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.081617 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c54f766-d384-46e5-a29f-2d1ac4181ba6-kube-api-access-4jdvj" (OuterVolumeSpecName: "kube-api-access-4jdvj") pod "0c54f766-d384-46e5-a29f-2d1ac4181ba6" (UID: "0c54f766-d384-46e5-a29f-2d1ac4181ba6"). InnerVolumeSpecName "kube-api-access-4jdvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.082067 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9de47e9d-5a4c-4563-b419-5ee9ca102f08" (UID: "9de47e9d-5a4c-4563-b419-5ee9ca102f08"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.084483 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-scripts" (OuterVolumeSpecName: "scripts") pod "0c54f766-d384-46e5-a29f-2d1ac4181ba6" (UID: "0c54f766-d384-46e5-a29f-2d1ac4181ba6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.113726 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-config-data" (OuterVolumeSpecName: "config-data") pod "0c54f766-d384-46e5-a29f-2d1ac4181ba6" (UID: "0c54f766-d384-46e5-a29f-2d1ac4181ba6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.116141 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-config-data" (OuterVolumeSpecName: "config-data") pod "9de47e9d-5a4c-4563-b419-5ee9ca102f08" (UID: "9de47e9d-5a4c-4563-b419-5ee9ca102f08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.176516 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.176826 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.176835 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.176843 4943 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c54f766-d384-46e5-a29f-2d1ac4181ba6-logs\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.176853 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9de47e9d-5a4c-4563-b419-5ee9ca102f08-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.176863 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.176874 4943 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9de47e9d-5a4c-4563-b419-5ee9ca102f08-logs\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.177384 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76jt5\" (UniqueName: \"kubernetes.io/projected/9de47e9d-5a4c-4563-b419-5ee9ca102f08-kube-api-access-76jt5\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.177437 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c54f766-d384-46e5-a29f-2d1ac4181ba6-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.177452 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jdvj\" (UniqueName: \"kubernetes.io/projected/0c54f766-d384-46e5-a29f-2d1ac4181ba6-kube-api-access-4jdvj\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.612935 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-1" event={"ID":"9de47e9d-5a4c-4563-b419-5ee9ca102f08","Type":"ContainerDied","Data":"2eebbd864927ecd09c680823d4d017b0e1118cfcb17e687dc26ed8babaa07459"} Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.612983 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-1" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.612990 4943 scope.go:117] "RemoveContainer" containerID="a5ecee7ed59649a14a5215ade747df8815db39dc12ef3a2ad5bc0fe6e1475653" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.615039 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-2" event={"ID":"0c54f766-d384-46e5-a29f-2d1ac4181ba6","Type":"ContainerDied","Data":"b734712b8f614a57185333054b5bc0b1d67b6be062141e9dced6db7b1e94a5bc"} Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.615123 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-2" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.633503 4943 scope.go:117] "RemoveContainer" containerID="4743f6093002a405c3c27f10ca16dd54e64b2b362666f710735fb3fc52c59e7a" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.648501 4943 scope.go:117] "RemoveContainer" containerID="1df6307739aed566ad0c5b08aaa675bc773c49be4f9c6c4e519561a97e23defb" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.655674 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.669680 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-2"] Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.670097 4943 scope.go:117] "RemoveContainer" containerID="a6ad19bb3ab132a00ade526b019e2fd89bffbfc25b38e8a321bec0af62d93f00" Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.677552 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 04 10:37:25 crc kubenswrapper[4943]: I1204 10:37:25.684534 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-1"] Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.400310 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-jm9vm"] Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.412856 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-jm9vm"] Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.422675 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.422965 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" containerName="cinder-scheduler" containerID="cri-o://89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975" gracePeriod=30 Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.423010 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" containerName="probe" containerID="cri-o://44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e" gracePeriod=30 Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.443303 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.444251 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="dcc96db6-b608-422f-a293-689b490741c5" containerName="cinder-backup" containerID="cri-o://30b3cdf6a5aed0a6e51a9d33ffd2692c4024394c53a586933ae10c02afd41e2d" gracePeriod=30 Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.447593 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="dcc96db6-b608-422f-a293-689b490741c5" containerName="probe" containerID="cri-o://ad98ac19f6f5811158d5e0064d662f1bd22e6182597088226f46659be1a2ead7" gracePeriod=30 Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.475969 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.488478 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.488812 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="99251963-b5ca-4bc8-a37b-e38d8f2f72db" containerName="cinder-api-log" containerID="cri-o://9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5" gracePeriod=30 Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.488980 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="99251963-b5ca-4bc8-a37b-e38d8f2f72db" containerName="cinder-api" containerID="cri-o://cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb" gracePeriod=30 Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.498861 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder2117-account-delete-nddrq"] Dec 04 10:37:26 crc kubenswrapper[4943]: E1204 10:37:26.499108 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" containerName="cinder-api-log" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.499124 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" containerName="cinder-api-log" Dec 04 10:37:26 crc kubenswrapper[4943]: E1204 10:37:26.499136 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" containerName="cinder-api" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.499144 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" containerName="cinder-api" Dec 04 10:37:26 crc kubenswrapper[4943]: E1204 10:37:26.499156 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" containerName="cinder-api" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.499162 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" containerName="cinder-api" Dec 04 10:37:26 crc kubenswrapper[4943]: E1204 10:37:26.499177 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" containerName="cinder-api-log" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.499182 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" containerName="cinder-api-log" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.499314 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" containerName="cinder-api-log" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.499327 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" containerName="cinder-api" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.499334 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" containerName="cinder-api" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.499344 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" containerName="cinder-api-log" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.500135 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.505716 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder2117-account-delete-nddrq"] Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.572662 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c54f766-d384-46e5-a29f-2d1ac4181ba6" path="/var/lib/kubelet/pods/0c54f766-d384-46e5-a29f-2d1ac4181ba6/volumes" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.573267 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9de47e9d-5a4c-4563-b419-5ee9ca102f08" path="/var/lib/kubelet/pods/9de47e9d-5a4c-4563-b419-5ee9ca102f08/volumes" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.573838 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3fd712b-e6d8-4282-9f14-9223a8e76bda" path="/var/lib/kubelet/pods/d3fd712b-e6d8-4282-9f14-9223a8e76bda/volumes" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.600887 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcrzh\" (UniqueName: \"kubernetes.io/projected/4912937b-2cf5-4103-b1fd-f6198f77de16-kube-api-access-xcrzh\") pod \"cinder2117-account-delete-nddrq\" (UID: \"4912937b-2cf5-4103-b1fd-f6198f77de16\") " pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.600952 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4912937b-2cf5-4103-b1fd-f6198f77de16-operator-scripts\") pod \"cinder2117-account-delete-nddrq\" (UID: \"4912937b-2cf5-4103-b1fd-f6198f77de16\") " pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.631236 4943 generic.go:334] "Generic (PLEG): container finished" podID="99251963-b5ca-4bc8-a37b-e38d8f2f72db" containerID="9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5" exitCode=143 Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.631323 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"99251963-b5ca-4bc8-a37b-e38d8f2f72db","Type":"ContainerDied","Data":"9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5"} Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.702354 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcrzh\" (UniqueName: \"kubernetes.io/projected/4912937b-2cf5-4103-b1fd-f6198f77de16-kube-api-access-xcrzh\") pod \"cinder2117-account-delete-nddrq\" (UID: \"4912937b-2cf5-4103-b1fd-f6198f77de16\") " pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.702435 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4912937b-2cf5-4103-b1fd-f6198f77de16-operator-scripts\") pod \"cinder2117-account-delete-nddrq\" (UID: \"4912937b-2cf5-4103-b1fd-f6198f77de16\") " pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.703342 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4912937b-2cf5-4103-b1fd-f6198f77de16-operator-scripts\") pod \"cinder2117-account-delete-nddrq\" (UID: \"4912937b-2cf5-4103-b1fd-f6198f77de16\") " pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.722448 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcrzh\" (UniqueName: \"kubernetes.io/projected/4912937b-2cf5-4103-b1fd-f6198f77de16-kube-api-access-xcrzh\") pod \"cinder2117-account-delete-nddrq\" (UID: \"4912937b-2cf5-4103-b1fd-f6198f77de16\") " pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.777748 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.820947 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.905815 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-machine-id\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.905874 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-iscsi\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.905908 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-config-data\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.905951 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-scripts\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.905969 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-nvme\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.905987 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-sys\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906001 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-locks-brick\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906021 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-lib-cinder\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906059 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-config-data-custom\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906104 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-run\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906141 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snrz9\" (UniqueName: \"kubernetes.io/projected/5dbc6cc7-a807-4f74-ad16-feecbab07846-kube-api-access-snrz9\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906157 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-dev\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906199 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-lib-modules\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906226 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-locks-cinder\") pod \"5dbc6cc7-a807-4f74-ad16-feecbab07846\" (UID: \"5dbc6cc7-a807-4f74-ad16-feecbab07846\") " Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906544 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906544 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906565 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906579 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906591 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-run" (OuterVolumeSpecName: "run") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.906658 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.907147 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.907243 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-sys" (OuterVolumeSpecName: "sys") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.907279 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.907277 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-dev" (OuterVolumeSpecName: "dev") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.910400 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-scripts" (OuterVolumeSpecName: "scripts") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.911150 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dbc6cc7-a807-4f74-ad16-feecbab07846-kube-api-access-snrz9" (OuterVolumeSpecName: "kube-api-access-snrz9") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "kube-api-access-snrz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.911321 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:26 crc kubenswrapper[4943]: I1204 10:37:26.985033 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-config-data" (OuterVolumeSpecName: "config-data") pod "5dbc6cc7-a807-4f74-ad16-feecbab07846" (UID: "5dbc6cc7-a807-4f74-ad16-feecbab07846"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.007999 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008037 4943 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008049 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008061 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008072 4943 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008083 4943 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-sys\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008095 4943 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008106 4943 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008117 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5dbc6cc7-a807-4f74-ad16-feecbab07846-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008129 4943 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-run\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008141 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snrz9\" (UniqueName: \"kubernetes.io/projected/5dbc6cc7-a807-4f74-ad16-feecbab07846-kube-api-access-snrz9\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008153 4943 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-dev\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008164 4943 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.008175 4943 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5dbc6cc7-a807-4f74-ad16-feecbab07846-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.057249 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder2117-account-delete-nddrq"] Dec 04 10:37:27 crc kubenswrapper[4943]: W1204 10:37:27.064329 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4912937b_2cf5_4103_b1fd_f6198f77de16.slice/crio-e2aa36467cd5f4c48829efa28d957cb748e3bb7a9630cfe88615d3c695f21615 WatchSource:0}: Error finding container e2aa36467cd5f4c48829efa28d957cb748e3bb7a9630cfe88615d3c695f21615: Status 404 returned error can't find the container with id e2aa36467cd5f4c48829efa28d957cb748e3bb7a9630cfe88615d3c695f21615 Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.525536 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.648765 4943 generic.go:334] "Generic (PLEG): container finished" podID="3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" containerID="44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e" exitCode=0 Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.649019 4943 generic.go:334] "Generic (PLEG): container finished" podID="3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" containerID="89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975" exitCode=0 Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.648831 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.648871 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c","Type":"ContainerDied","Data":"44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e"} Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.649395 4943 scope.go:117] "RemoveContainer" containerID="44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.649345 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c","Type":"ContainerDied","Data":"89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975"} Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.649740 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c","Type":"ContainerDied","Data":"1c3a5773f69bf227bd2c5162dfc0fd97fb4b3dcc34f4cab0a6bff48b8f12d7bb"} Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.653669 4943 generic.go:334] "Generic (PLEG): container finished" podID="dcc96db6-b608-422f-a293-689b490741c5" containerID="ad98ac19f6f5811158d5e0064d662f1bd22e6182597088226f46659be1a2ead7" exitCode=0 Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.653716 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"dcc96db6-b608-422f-a293-689b490741c5","Type":"ContainerDied","Data":"ad98ac19f6f5811158d5e0064d662f1bd22e6182597088226f46659be1a2ead7"} Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.656143 4943 generic.go:334] "Generic (PLEG): container finished" podID="4912937b-2cf5-4103-b1fd-f6198f77de16" containerID="af56e0e6966666528f27742b897bea116befeb238c729c0b5ba875f4ecffea46" exitCode=0 Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.656653 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" event={"ID":"4912937b-2cf5-4103-b1fd-f6198f77de16","Type":"ContainerDied","Data":"af56e0e6966666528f27742b897bea116befeb238c729c0b5ba875f4ecffea46"} Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.656696 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" event={"ID":"4912937b-2cf5-4103-b1fd-f6198f77de16","Type":"ContainerStarted","Data":"e2aa36467cd5f4c48829efa28d957cb748e3bb7a9630cfe88615d3c695f21615"} Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.659607 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"5dbc6cc7-a807-4f74-ad16-feecbab07846","Type":"ContainerDied","Data":"b624f3c16d5991d677207b596e4e0beb3df23e66d031449746484c7fad83d725"} Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.659685 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.693448 4943 scope.go:117] "RemoveContainer" containerID="89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.705452 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.716890 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-config-data-custom\") pod \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.717002 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-scripts\") pod \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.717134 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-config-data\") pod \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.717175 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kq6pd\" (UniqueName: \"kubernetes.io/projected/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-kube-api-access-kq6pd\") pod \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.717335 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-etc-machine-id\") pod \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\" (UID: \"3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c\") " Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.717473 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.717632 4943 scope.go:117] "RemoveContainer" containerID="44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.718817 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" (UID: "3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:27 crc kubenswrapper[4943]: E1204 10:37:27.719442 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e\": container with ID starting with 44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e not found: ID does not exist" containerID="44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.719481 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e"} err="failed to get container status \"44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e\": rpc error: code = NotFound desc = could not find container \"44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e\": container with ID starting with 44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e not found: ID does not exist" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.719510 4943 scope.go:117] "RemoveContainer" containerID="89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975" Dec 04 10:37:27 crc kubenswrapper[4943]: E1204 10:37:27.720149 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975\": container with ID starting with 89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975 not found: ID does not exist" containerID="89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.720240 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975"} err="failed to get container status \"89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975\": rpc error: code = NotFound desc = could not find container \"89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975\": container with ID starting with 89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975 not found: ID does not exist" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.720285 4943 scope.go:117] "RemoveContainer" containerID="44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.720685 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e"} err="failed to get container status \"44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e\": rpc error: code = NotFound desc = could not find container \"44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e\": container with ID starting with 44d7cf9d6f025fbeb18d418c49121b348a901112d32c607aca00054601c1c82e not found: ID does not exist" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.720715 4943 scope.go:117] "RemoveContainer" containerID="89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.721328 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975"} err="failed to get container status \"89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975\": rpc error: code = NotFound desc = could not find container \"89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975\": container with ID starting with 89414b48f34726503ab80fc50a8de403050745b95d2bbad42623c861daf37975 not found: ID does not exist" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.721391 4943 scope.go:117] "RemoveContainer" containerID="a5e05e77639a143c450e901da7187ae9ef9fe11a14e7b0d28d9484b7f49c0a68" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.722869 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" (UID: "3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.723478 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-scripts" (OuterVolumeSpecName: "scripts") pod "3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" (UID: "3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.725715 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-kube-api-access-kq6pd" (OuterVolumeSpecName: "kube-api-access-kq6pd") pod "3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" (UID: "3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c"). InnerVolumeSpecName "kube-api-access-kq6pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.788373 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-config-data" (OuterVolumeSpecName: "config-data") pod "3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" (UID: "3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.819438 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.819476 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kq6pd\" (UniqueName: \"kubernetes.io/projected/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-kube-api-access-kq6pd\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.819486 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.819498 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.819506 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.841608 4943 scope.go:117] "RemoveContainer" containerID="cdb78185274d9ca800f5449ab5d23529653eb05dfc50d6507d84c36a3415a769" Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.980963 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 04 10:37:27 crc kubenswrapper[4943]: I1204 10:37:27.988200 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 04 10:37:28 crc kubenswrapper[4943]: I1204 10:37:28.584591 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" path="/var/lib/kubelet/pods/3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c/volumes" Dec 04 10:37:28 crc kubenswrapper[4943]: I1204 10:37:28.585960 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" path="/var/lib/kubelet/pods/5dbc6cc7-a807-4f74-ad16-feecbab07846/volumes" Dec 04 10:37:28 crc kubenswrapper[4943]: I1204 10:37:28.883753 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" Dec 04 10:37:29 crc kubenswrapper[4943]: I1204 10:37:29.037448 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcrzh\" (UniqueName: \"kubernetes.io/projected/4912937b-2cf5-4103-b1fd-f6198f77de16-kube-api-access-xcrzh\") pod \"4912937b-2cf5-4103-b1fd-f6198f77de16\" (UID: \"4912937b-2cf5-4103-b1fd-f6198f77de16\") " Dec 04 10:37:29 crc kubenswrapper[4943]: I1204 10:37:29.037593 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4912937b-2cf5-4103-b1fd-f6198f77de16-operator-scripts\") pod \"4912937b-2cf5-4103-b1fd-f6198f77de16\" (UID: \"4912937b-2cf5-4103-b1fd-f6198f77de16\") " Dec 04 10:37:29 crc kubenswrapper[4943]: I1204 10:37:29.038379 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4912937b-2cf5-4103-b1fd-f6198f77de16-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4912937b-2cf5-4103-b1fd-f6198f77de16" (UID: "4912937b-2cf5-4103-b1fd-f6198f77de16"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:37:29 crc kubenswrapper[4943]: I1204 10:37:29.041764 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4912937b-2cf5-4103-b1fd-f6198f77de16-kube-api-access-xcrzh" (OuterVolumeSpecName: "kube-api-access-xcrzh") pod "4912937b-2cf5-4103-b1fd-f6198f77de16" (UID: "4912937b-2cf5-4103-b1fd-f6198f77de16"). InnerVolumeSpecName "kube-api-access-xcrzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:29 crc kubenswrapper[4943]: I1204 10:37:29.138944 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4912937b-2cf5-4103-b1fd-f6198f77de16-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:29 crc kubenswrapper[4943]: I1204 10:37:29.138993 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcrzh\" (UniqueName: \"kubernetes.io/projected/4912937b-2cf5-4103-b1fd-f6198f77de16-kube-api-access-xcrzh\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:29 crc kubenswrapper[4943]: I1204 10:37:29.647320 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-0" podUID="99251963-b5ca-4bc8-a37b-e38d8f2f72db" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.97:8776/healthcheck\": read tcp 10.217.0.2:40836->10.217.0.97:8776: read: connection reset by peer" Dec 04 10:37:29 crc kubenswrapper[4943]: I1204 10:37:29.681454 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" event={"ID":"4912937b-2cf5-4103-b1fd-f6198f77de16","Type":"ContainerDied","Data":"e2aa36467cd5f4c48829efa28d957cb748e3bb7a9630cfe88615d3c695f21615"} Dec 04 10:37:29 crc kubenswrapper[4943]: I1204 10:37:29.681747 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2aa36467cd5f4c48829efa28d957cb748e3bb7a9630cfe88615d3c695f21615" Dec 04 10:37:29 crc kubenswrapper[4943]: I1204 10:37:29.681526 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder2117-account-delete-nddrq" Dec 04 10:37:29 crc kubenswrapper[4943]: I1204 10:37:29.971445 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.161667 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv7wg\" (UniqueName: \"kubernetes.io/projected/99251963-b5ca-4bc8-a37b-e38d8f2f72db-kube-api-access-bv7wg\") pod \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.161755 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-scripts\") pod \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.161803 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99251963-b5ca-4bc8-a37b-e38d8f2f72db-etc-machine-id\") pod \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.161881 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-config-data\") pod \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.161971 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99251963-b5ca-4bc8-a37b-e38d8f2f72db-logs\") pod \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.162150 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/99251963-b5ca-4bc8-a37b-e38d8f2f72db-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "99251963-b5ca-4bc8-a37b-e38d8f2f72db" (UID: "99251963-b5ca-4bc8-a37b-e38d8f2f72db"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.162326 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-config-data-custom\") pod \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\" (UID: \"99251963-b5ca-4bc8-a37b-e38d8f2f72db\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.162678 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/99251963-b5ca-4bc8-a37b-e38d8f2f72db-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.162838 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99251963-b5ca-4bc8-a37b-e38d8f2f72db-logs" (OuterVolumeSpecName: "logs") pod "99251963-b5ca-4bc8-a37b-e38d8f2f72db" (UID: "99251963-b5ca-4bc8-a37b-e38d8f2f72db"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.170498 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "99251963-b5ca-4bc8-a37b-e38d8f2f72db" (UID: "99251963-b5ca-4bc8-a37b-e38d8f2f72db"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.170531 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-scripts" (OuterVolumeSpecName: "scripts") pod "99251963-b5ca-4bc8-a37b-e38d8f2f72db" (UID: "99251963-b5ca-4bc8-a37b-e38d8f2f72db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.170569 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99251963-b5ca-4bc8-a37b-e38d8f2f72db-kube-api-access-bv7wg" (OuterVolumeSpecName: "kube-api-access-bv7wg") pod "99251963-b5ca-4bc8-a37b-e38d8f2f72db" (UID: "99251963-b5ca-4bc8-a37b-e38d8f2f72db"). InnerVolumeSpecName "kube-api-access-bv7wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.209973 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-config-data" (OuterVolumeSpecName: "config-data") pod "99251963-b5ca-4bc8-a37b-e38d8f2f72db" (UID: "99251963-b5ca-4bc8-a37b-e38d8f2f72db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.264859 4943 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99251963-b5ca-4bc8-a37b-e38d8f2f72db-logs\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.264908 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.264921 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv7wg\" (UniqueName: \"kubernetes.io/projected/99251963-b5ca-4bc8-a37b-e38d8f2f72db-kube-api-access-bv7wg\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.264932 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.264942 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99251963-b5ca-4bc8-a37b-e38d8f2f72db-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.692225 4943 generic.go:334] "Generic (PLEG): container finished" podID="99251963-b5ca-4bc8-a37b-e38d8f2f72db" containerID="cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb" exitCode=0 Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.692265 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.692292 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"99251963-b5ca-4bc8-a37b-e38d8f2f72db","Type":"ContainerDied","Data":"cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb"} Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.692332 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"99251963-b5ca-4bc8-a37b-e38d8f2f72db","Type":"ContainerDied","Data":"fe5d5c271c5499d9e3d87ff1aa172c1ca1deea9242ca3384fff35337092152ce"} Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.692352 4943 scope.go:117] "RemoveContainer" containerID="cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.695153 4943 generic.go:334] "Generic (PLEG): container finished" podID="dcc96db6-b608-422f-a293-689b490741c5" containerID="30b3cdf6a5aed0a6e51a9d33ffd2692c4024394c53a586933ae10c02afd41e2d" exitCode=0 Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.695182 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"dcc96db6-b608-422f-a293-689b490741c5","Type":"ContainerDied","Data":"30b3cdf6a5aed0a6e51a9d33ffd2692c4024394c53a586933ae10c02afd41e2d"} Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.720567 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.726046 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.742999 4943 scope.go:117] "RemoveContainer" containerID="9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.770626 4943 scope.go:117] "RemoveContainer" containerID="cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb" Dec 04 10:37:30 crc kubenswrapper[4943]: E1204 10:37:30.771239 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb\": container with ID starting with cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb not found: ID does not exist" containerID="cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.771272 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb"} err="failed to get container status \"cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb\": rpc error: code = NotFound desc = could not find container \"cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb\": container with ID starting with cf1be05a31d3d58ed4d6e8291357f83855273e14a8601d50a7cc185cfc0a01cb not found: ID does not exist" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.771394 4943 scope.go:117] "RemoveContainer" containerID="9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5" Dec 04 10:37:30 crc kubenswrapper[4943]: E1204 10:37:30.771684 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5\": container with ID starting with 9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5 not found: ID does not exist" containerID="9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.771733 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5"} err="failed to get container status \"9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5\": rpc error: code = NotFound desc = could not find container \"9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5\": container with ID starting with 9f12e30750d265635e2c9d1ddecbb798b998b225375c5889eeec1762f1d8b6d5 not found: ID does not exist" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.869510 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.881551 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-lib-cinder\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.881706 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.882067 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-dev" (OuterVolumeSpecName: "dev") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.882283 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-dev\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.882509 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-scripts\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.883158 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-config-data\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.883903 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-sys" (OuterVolumeSpecName: "sys") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.884246 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-sys\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.884442 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-locks-cinder\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.884617 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.884891 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fglr\" (UniqueName: \"kubernetes.io/projected/dcc96db6-b608-422f-a293-689b490741c5-kube-api-access-7fglr\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.885303 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-config-data-custom\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.885508 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-scripts" (OuterVolumeSpecName: "scripts") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.885652 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.885789 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-machine-id\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.885974 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-nvme\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.886019 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.886343 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.886315 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-lib-modules\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.886949 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.887284 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-iscsi\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.888177 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-run" (OuterVolumeSpecName: "run") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.888158 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-run\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.888917 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.889077 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.889181 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-locks-brick\") pod \"dcc96db6-b608-422f-a293-689b490741c5\" (UID: \"dcc96db6-b608-422f-a293-689b490741c5\") " Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.890449 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.892542 4943 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.892736 4943 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.892940 4943 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.893099 4943 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-run\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.893276 4943 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.893926 4943 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.895323 4943 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-dev\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.895353 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.890475 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcc96db6-b608-422f-a293-689b490741c5-kube-api-access-7fglr" (OuterVolumeSpecName: "kube-api-access-7fglr") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "kube-api-access-7fglr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.895364 4943 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-sys\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.895426 4943 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/dcc96db6-b608-422f-a293-689b490741c5-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.895444 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.951349 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-config-data" (OuterVolumeSpecName: "config-data") pod "dcc96db6-b608-422f-a293-689b490741c5" (UID: "dcc96db6-b608-422f-a293-689b490741c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.996740 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcc96db6-b608-422f-a293-689b490741c5-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:30 crc kubenswrapper[4943]: I1204 10:37:30.996779 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fglr\" (UniqueName: \"kubernetes.io/projected/dcc96db6-b608-422f-a293-689b490741c5-kube-api-access-7fglr\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.507384 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-k85jz"] Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.543296 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-k85jz"] Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.551600 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder2117-account-delete-nddrq"] Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.563015 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5"] Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.576279 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder2117-account-delete-nddrq"] Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.582061 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-2117-account-create-update-vzqx5"] Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.688813 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-create-d476h"] Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689159 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689186 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689202 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689226 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689237 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689246 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689257 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="cinder-volume" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689267 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="cinder-volume" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689280 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" containerName="cinder-scheduler" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689289 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" containerName="cinder-scheduler" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689300 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcc96db6-b608-422f-a293-689b490741c5" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689307 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcc96db6-b608-422f-a293-689b490741c5" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689328 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99251963-b5ca-4bc8-a37b-e38d8f2f72db" containerName="cinder-api" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689336 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="99251963-b5ca-4bc8-a37b-e38d8f2f72db" containerName="cinder-api" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689352 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcc96db6-b608-422f-a293-689b490741c5" containerName="cinder-backup" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689359 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcc96db6-b608-422f-a293-689b490741c5" containerName="cinder-backup" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689371 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689378 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689390 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="cinder-volume" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689398 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="cinder-volume" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689407 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4912937b-2cf5-4103-b1fd-f6198f77de16" containerName="mariadb-account-delete" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689414 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="4912937b-2cf5-4103-b1fd-f6198f77de16" containerName="mariadb-account-delete" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689423 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="cinder-volume" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689430 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="cinder-volume" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.689444 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99251963-b5ca-4bc8-a37b-e38d8f2f72db" containerName="cinder-api-log" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689452 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="99251963-b5ca-4bc8-a37b-e38d8f2f72db" containerName="cinder-api-log" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689594 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689610 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689621 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="cinder-volume" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689631 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="cinder-volume" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689639 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcc96db6-b608-422f-a293-689b490741c5" containerName="cinder-backup" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689649 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcc96db6-b608-422f-a293-689b490741c5" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689662 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" containerName="cinder-scheduler" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689673 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="4912937b-2cf5-4103-b1fd-f6198f77de16" containerName="mariadb-account-delete" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689688 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689698 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="99251963-b5ca-4bc8-a37b-e38d8f2f72db" containerName="cinder-api" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689707 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d36408d-5a2b-426a-a5b3-2b1bb0d30d2c" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.689719 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="99251963-b5ca-4bc8-a37b-e38d8f2f72db" containerName="cinder-api-log" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.759438 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-d476h" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.769585 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv"] Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.770016 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.770033 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: E1204 10:37:31.770068 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="cinder-volume" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.770077 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="cinder-volume" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.770238 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="probe" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.770261 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="cinder-volume" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.770270 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dbc6cc7-a807-4f74-ad16-feecbab07846" containerName="cinder-volume" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.770932 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.776839 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-d476h"] Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.777335 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-db-secret" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.783714 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.800983 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"dcc96db6-b608-422f-a293-689b490741c5","Type":"ContainerDied","Data":"2a71f9dcf8d9544e29935cf187f26cf0e141b45157ec9949221e1c7101ab4f04"} Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.801137 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv"] Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.801254 4943 scope.go:117] "RemoveContainer" containerID="ad98ac19f6f5811158d5e0064d662f1bd22e6182597088226f46659be1a2ead7" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.839507 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.853707 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.858433 4943 scope.go:117] "RemoveContainer" containerID="30b3cdf6a5aed0a6e51a9d33ffd2692c4024394c53a586933ae10c02afd41e2d" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.860018 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c80881c-bab1-4465-82f0-dcedc10fe88c-operator-scripts\") pod \"cinder-6acd-account-create-update-ts5vv\" (UID: \"8c80881c-bab1-4465-82f0-dcedc10fe88c\") " pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.860107 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1565f2bc-590f-4b50-ab58-05565b6df99c-operator-scripts\") pod \"cinder-db-create-d476h\" (UID: \"1565f2bc-590f-4b50-ab58-05565b6df99c\") " pod="cinder-kuttl-tests/cinder-db-create-d476h" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.860153 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4p4c\" (UniqueName: \"kubernetes.io/projected/8c80881c-bab1-4465-82f0-dcedc10fe88c-kube-api-access-x4p4c\") pod \"cinder-6acd-account-create-update-ts5vv\" (UID: \"8c80881c-bab1-4465-82f0-dcedc10fe88c\") " pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.860385 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x86p\" (UniqueName: \"kubernetes.io/projected/1565f2bc-590f-4b50-ab58-05565b6df99c-kube-api-access-8x86p\") pod \"cinder-db-create-d476h\" (UID: \"1565f2bc-590f-4b50-ab58-05565b6df99c\") " pod="cinder-kuttl-tests/cinder-db-create-d476h" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.962430 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1565f2bc-590f-4b50-ab58-05565b6df99c-operator-scripts\") pod \"cinder-db-create-d476h\" (UID: \"1565f2bc-590f-4b50-ab58-05565b6df99c\") " pod="cinder-kuttl-tests/cinder-db-create-d476h" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.962508 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4p4c\" (UniqueName: \"kubernetes.io/projected/8c80881c-bab1-4465-82f0-dcedc10fe88c-kube-api-access-x4p4c\") pod \"cinder-6acd-account-create-update-ts5vv\" (UID: \"8c80881c-bab1-4465-82f0-dcedc10fe88c\") " pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.962549 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x86p\" (UniqueName: \"kubernetes.io/projected/1565f2bc-590f-4b50-ab58-05565b6df99c-kube-api-access-8x86p\") pod \"cinder-db-create-d476h\" (UID: \"1565f2bc-590f-4b50-ab58-05565b6df99c\") " pod="cinder-kuttl-tests/cinder-db-create-d476h" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.962571 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c80881c-bab1-4465-82f0-dcedc10fe88c-operator-scripts\") pod \"cinder-6acd-account-create-update-ts5vv\" (UID: \"8c80881c-bab1-4465-82f0-dcedc10fe88c\") " pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.963341 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c80881c-bab1-4465-82f0-dcedc10fe88c-operator-scripts\") pod \"cinder-6acd-account-create-update-ts5vv\" (UID: \"8c80881c-bab1-4465-82f0-dcedc10fe88c\") " pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.963754 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1565f2bc-590f-4b50-ab58-05565b6df99c-operator-scripts\") pod \"cinder-db-create-d476h\" (UID: \"1565f2bc-590f-4b50-ab58-05565b6df99c\") " pod="cinder-kuttl-tests/cinder-db-create-d476h" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.980772 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4p4c\" (UniqueName: \"kubernetes.io/projected/8c80881c-bab1-4465-82f0-dcedc10fe88c-kube-api-access-x4p4c\") pod \"cinder-6acd-account-create-update-ts5vv\" (UID: \"8c80881c-bab1-4465-82f0-dcedc10fe88c\") " pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" Dec 04 10:37:31 crc kubenswrapper[4943]: I1204 10:37:31.980853 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x86p\" (UniqueName: \"kubernetes.io/projected/1565f2bc-590f-4b50-ab58-05565b6df99c-kube-api-access-8x86p\") pod \"cinder-db-create-d476h\" (UID: \"1565f2bc-590f-4b50-ab58-05565b6df99c\") " pod="cinder-kuttl-tests/cinder-db-create-d476h" Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.117699 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-d476h" Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.135901 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.343138 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-d476h"] Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.576832 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="121f8a1e-5d27-4f26-9578-65f06fd83846" path="/var/lib/kubelet/pods/121f8a1e-5d27-4f26-9578-65f06fd83846/volumes" Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.577782 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4912937b-2cf5-4103-b1fd-f6198f77de16" path="/var/lib/kubelet/pods/4912937b-2cf5-4103-b1fd-f6198f77de16/volumes" Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.578405 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99251963-b5ca-4bc8-a37b-e38d8f2f72db" path="/var/lib/kubelet/pods/99251963-b5ca-4bc8-a37b-e38d8f2f72db/volumes" Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.579581 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bc2510f-d4d9-46dd-b5fc-409e89ababbd" path="/var/lib/kubelet/pods/9bc2510f-d4d9-46dd-b5fc-409e89ababbd/volumes" Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.580060 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcc96db6-b608-422f-a293-689b490741c5" path="/var/lib/kubelet/pods/dcc96db6-b608-422f-a293-689b490741c5/volumes" Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.610380 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv"] Dec 04 10:37:32 crc kubenswrapper[4943]: W1204 10:37:32.614475 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c80881c_bab1_4465_82f0_dcedc10fe88c.slice/crio-333ca2c0fc7f67db86dce5b0256a24906d289b24811f280a41b851d122c1f142 WatchSource:0}: Error finding container 333ca2c0fc7f67db86dce5b0256a24906d289b24811f280a41b851d122c1f142: Status 404 returned error can't find the container with id 333ca2c0fc7f67db86dce5b0256a24906d289b24811f280a41b851d122c1f142 Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.793714 4943 generic.go:334] "Generic (PLEG): container finished" podID="1565f2bc-590f-4b50-ab58-05565b6df99c" containerID="9716c02c743af52efd8e9b59789dfbb5f003a14d06f53dcf23af6ca00f6ad0ec" exitCode=0 Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.793770 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-d476h" event={"ID":"1565f2bc-590f-4b50-ab58-05565b6df99c","Type":"ContainerDied","Data":"9716c02c743af52efd8e9b59789dfbb5f003a14d06f53dcf23af6ca00f6ad0ec"} Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.793826 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-d476h" event={"ID":"1565f2bc-590f-4b50-ab58-05565b6df99c","Type":"ContainerStarted","Data":"2218857fa32ef1ca8f8685d5f749fe857f2b8b8dc9bf839c54c019b91642813c"} Dec 04 10:37:32 crc kubenswrapper[4943]: I1204 10:37:32.798309 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" event={"ID":"8c80881c-bab1-4465-82f0-dcedc10fe88c","Type":"ContainerStarted","Data":"333ca2c0fc7f67db86dce5b0256a24906d289b24811f280a41b851d122c1f142"} Dec 04 10:37:33 crc kubenswrapper[4943]: I1204 10:37:33.820417 4943 generic.go:334] "Generic (PLEG): container finished" podID="8c80881c-bab1-4465-82f0-dcedc10fe88c" containerID="6066e7ec929e202bc35aa6fe00dabaa4bb894b0e31f40d0151ca84aed43b0171" exitCode=0 Dec 04 10:37:33 crc kubenswrapper[4943]: I1204 10:37:33.821607 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" event={"ID":"8c80881c-bab1-4465-82f0-dcedc10fe88c","Type":"ContainerDied","Data":"6066e7ec929e202bc35aa6fe00dabaa4bb894b0e31f40d0151ca84aed43b0171"} Dec 04 10:37:34 crc kubenswrapper[4943]: I1204 10:37:34.144868 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-d476h" Dec 04 10:37:34 crc kubenswrapper[4943]: I1204 10:37:34.302950 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x86p\" (UniqueName: \"kubernetes.io/projected/1565f2bc-590f-4b50-ab58-05565b6df99c-kube-api-access-8x86p\") pod \"1565f2bc-590f-4b50-ab58-05565b6df99c\" (UID: \"1565f2bc-590f-4b50-ab58-05565b6df99c\") " Dec 04 10:37:34 crc kubenswrapper[4943]: I1204 10:37:34.303248 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1565f2bc-590f-4b50-ab58-05565b6df99c-operator-scripts\") pod \"1565f2bc-590f-4b50-ab58-05565b6df99c\" (UID: \"1565f2bc-590f-4b50-ab58-05565b6df99c\") " Dec 04 10:37:34 crc kubenswrapper[4943]: I1204 10:37:34.304480 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1565f2bc-590f-4b50-ab58-05565b6df99c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1565f2bc-590f-4b50-ab58-05565b6df99c" (UID: "1565f2bc-590f-4b50-ab58-05565b6df99c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:37:34 crc kubenswrapper[4943]: I1204 10:37:34.308887 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1565f2bc-590f-4b50-ab58-05565b6df99c-kube-api-access-8x86p" (OuterVolumeSpecName: "kube-api-access-8x86p") pod "1565f2bc-590f-4b50-ab58-05565b6df99c" (UID: "1565f2bc-590f-4b50-ab58-05565b6df99c"). InnerVolumeSpecName "kube-api-access-8x86p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:34 crc kubenswrapper[4943]: I1204 10:37:34.404737 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1565f2bc-590f-4b50-ab58-05565b6df99c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:34 crc kubenswrapper[4943]: I1204 10:37:34.404782 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x86p\" (UniqueName: \"kubernetes.io/projected/1565f2bc-590f-4b50-ab58-05565b6df99c-kube-api-access-8x86p\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:34 crc kubenswrapper[4943]: I1204 10:37:34.833467 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-d476h" Dec 04 10:37:34 crc kubenswrapper[4943]: I1204 10:37:34.833668 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-d476h" event={"ID":"1565f2bc-590f-4b50-ab58-05565b6df99c","Type":"ContainerDied","Data":"2218857fa32ef1ca8f8685d5f749fe857f2b8b8dc9bf839c54c019b91642813c"} Dec 04 10:37:34 crc kubenswrapper[4943]: I1204 10:37:34.833694 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2218857fa32ef1ca8f8685d5f749fe857f2b8b8dc9bf839c54c019b91642813c" Dec 04 10:37:35 crc kubenswrapper[4943]: I1204 10:37:35.119942 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" Dec 04 10:37:35 crc kubenswrapper[4943]: I1204 10:37:35.217674 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c80881c-bab1-4465-82f0-dcedc10fe88c-operator-scripts\") pod \"8c80881c-bab1-4465-82f0-dcedc10fe88c\" (UID: \"8c80881c-bab1-4465-82f0-dcedc10fe88c\") " Dec 04 10:37:35 crc kubenswrapper[4943]: I1204 10:37:35.217770 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4p4c\" (UniqueName: \"kubernetes.io/projected/8c80881c-bab1-4465-82f0-dcedc10fe88c-kube-api-access-x4p4c\") pod \"8c80881c-bab1-4465-82f0-dcedc10fe88c\" (UID: \"8c80881c-bab1-4465-82f0-dcedc10fe88c\") " Dec 04 10:37:35 crc kubenswrapper[4943]: I1204 10:37:35.218744 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c80881c-bab1-4465-82f0-dcedc10fe88c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8c80881c-bab1-4465-82f0-dcedc10fe88c" (UID: "8c80881c-bab1-4465-82f0-dcedc10fe88c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:37:35 crc kubenswrapper[4943]: I1204 10:37:35.224653 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c80881c-bab1-4465-82f0-dcedc10fe88c-kube-api-access-x4p4c" (OuterVolumeSpecName: "kube-api-access-x4p4c") pod "8c80881c-bab1-4465-82f0-dcedc10fe88c" (UID: "8c80881c-bab1-4465-82f0-dcedc10fe88c"). InnerVolumeSpecName "kube-api-access-x4p4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:35 crc kubenswrapper[4943]: I1204 10:37:35.318941 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8c80881c-bab1-4465-82f0-dcedc10fe88c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:35 crc kubenswrapper[4943]: I1204 10:37:35.318978 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4p4c\" (UniqueName: \"kubernetes.io/projected/8c80881c-bab1-4465-82f0-dcedc10fe88c-kube-api-access-x4p4c\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:35 crc kubenswrapper[4943]: I1204 10:37:35.846153 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" event={"ID":"8c80881c-bab1-4465-82f0-dcedc10fe88c","Type":"ContainerDied","Data":"333ca2c0fc7f67db86dce5b0256a24906d289b24811f280a41b851d122c1f142"} Dec 04 10:37:35 crc kubenswrapper[4943]: I1204 10:37:35.846557 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="333ca2c0fc7f67db86dce5b0256a24906d289b24811f280a41b851d122c1f142" Dec 04 10:37:35 crc kubenswrapper[4943]: I1204 10:37:35.846242 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv" Dec 04 10:37:36 crc kubenswrapper[4943]: I1204 10:37:36.937986 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-6kxff"] Dec 04 10:37:36 crc kubenswrapper[4943]: E1204 10:37:36.939563 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1565f2bc-590f-4b50-ab58-05565b6df99c" containerName="mariadb-database-create" Dec 04 10:37:36 crc kubenswrapper[4943]: I1204 10:37:36.939658 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1565f2bc-590f-4b50-ab58-05565b6df99c" containerName="mariadb-database-create" Dec 04 10:37:36 crc kubenswrapper[4943]: E1204 10:37:36.939727 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c80881c-bab1-4465-82f0-dcedc10fe88c" containerName="mariadb-account-create-update" Dec 04 10:37:36 crc kubenswrapper[4943]: I1204 10:37:36.939782 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c80881c-bab1-4465-82f0-dcedc10fe88c" containerName="mariadb-account-create-update" Dec 04 10:37:36 crc kubenswrapper[4943]: I1204 10:37:36.940013 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1565f2bc-590f-4b50-ab58-05565b6df99c" containerName="mariadb-database-create" Dec 04 10:37:36 crc kubenswrapper[4943]: I1204 10:37:36.940087 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c80881c-bab1-4465-82f0-dcedc10fe88c" containerName="mariadb-account-create-update" Dec 04 10:37:36 crc kubenswrapper[4943]: I1204 10:37:36.940666 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:36 crc kubenswrapper[4943]: I1204 10:37:36.942725 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Dec 04 10:37:36 crc kubenswrapper[4943]: I1204 10:37:36.943282 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Dec 04 10:37:36 crc kubenswrapper[4943]: I1204 10:37:36.945748 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"combined-ca-bundle" Dec 04 10:37:36 crc kubenswrapper[4943]: I1204 10:37:36.947348 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-6kxff"] Dec 04 10:37:36 crc kubenswrapper[4943]: I1204 10:37:36.949830 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-kf298" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.047128 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-db-sync-config-data\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.047170 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-combined-ca-bundle\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.047217 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-scripts\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.047238 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78919be1-5911-43e4-aa83-b8b06c443f4f-etc-machine-id\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.047522 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-config-data\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.047615 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4dbl\" (UniqueName: \"kubernetes.io/projected/78919be1-5911-43e4-aa83-b8b06c443f4f-kube-api-access-c4dbl\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.148671 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-config-data\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.148987 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4dbl\" (UniqueName: \"kubernetes.io/projected/78919be1-5911-43e4-aa83-b8b06c443f4f-kube-api-access-c4dbl\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.149139 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-db-sync-config-data\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.149255 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-combined-ca-bundle\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.149371 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-scripts\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.149492 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78919be1-5911-43e4-aa83-b8b06c443f4f-etc-machine-id\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.149601 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78919be1-5911-43e4-aa83-b8b06c443f4f-etc-machine-id\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.153740 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-scripts\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.153748 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-db-sync-config-data\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.154002 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-combined-ca-bundle\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.154455 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-config-data\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.169781 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4dbl\" (UniqueName: \"kubernetes.io/projected/78919be1-5911-43e4-aa83-b8b06c443f4f-kube-api-access-c4dbl\") pod \"cinder-db-sync-6kxff\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.305633 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.743614 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-6kxff"] Dec 04 10:37:37 crc kubenswrapper[4943]: I1204 10:37:37.867685 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-6kxff" event={"ID":"78919be1-5911-43e4-aa83-b8b06c443f4f","Type":"ContainerStarted","Data":"31e0aeda0a452bf1a43c7de3f14fd54f6b9408104301ded7b771fa4998729cd9"} Dec 04 10:37:38 crc kubenswrapper[4943]: I1204 10:37:38.876286 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-6kxff" event={"ID":"78919be1-5911-43e4-aa83-b8b06c443f4f","Type":"ContainerStarted","Data":"1f4ea23318021a79c31454ef74765c616fe201b11bff0faac5fcea5243e56dcc"} Dec 04 10:37:38 crc kubenswrapper[4943]: I1204 10:37:38.908239 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-db-sync-6kxff" podStartSLOduration=2.908218539 podStartE2EDuration="2.908218539s" podCreationTimestamp="2025-12-04 10:37:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:37:38.905431953 +0000 UTC m=+1307.494407871" watchObservedRunningTime="2025-12-04 10:37:38.908218539 +0000 UTC m=+1307.497194407" Dec 04 10:37:40 crc kubenswrapper[4943]: I1204 10:37:40.898386 4943 generic.go:334] "Generic (PLEG): container finished" podID="78919be1-5911-43e4-aa83-b8b06c443f4f" containerID="1f4ea23318021a79c31454ef74765c616fe201b11bff0faac5fcea5243e56dcc" exitCode=0 Dec 04 10:37:40 crc kubenswrapper[4943]: I1204 10:37:40.898518 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-6kxff" event={"ID":"78919be1-5911-43e4-aa83-b8b06c443f4f","Type":"ContainerDied","Data":"1f4ea23318021a79c31454ef74765c616fe201b11bff0faac5fcea5243e56dcc"} Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.184605 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.336778 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-db-sync-config-data\") pod \"78919be1-5911-43e4-aa83-b8b06c443f4f\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.336850 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-config-data\") pod \"78919be1-5911-43e4-aa83-b8b06c443f4f\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.336886 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4dbl\" (UniqueName: \"kubernetes.io/projected/78919be1-5911-43e4-aa83-b8b06c443f4f-kube-api-access-c4dbl\") pod \"78919be1-5911-43e4-aa83-b8b06c443f4f\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.336976 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-combined-ca-bundle\") pod \"78919be1-5911-43e4-aa83-b8b06c443f4f\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.337023 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78919be1-5911-43e4-aa83-b8b06c443f4f-etc-machine-id\") pod \"78919be1-5911-43e4-aa83-b8b06c443f4f\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.337056 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-scripts\") pod \"78919be1-5911-43e4-aa83-b8b06c443f4f\" (UID: \"78919be1-5911-43e4-aa83-b8b06c443f4f\") " Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.337265 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/78919be1-5911-43e4-aa83-b8b06c443f4f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "78919be1-5911-43e4-aa83-b8b06c443f4f" (UID: "78919be1-5911-43e4-aa83-b8b06c443f4f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.337404 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/78919be1-5911-43e4-aa83-b8b06c443f4f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.341960 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-scripts" (OuterVolumeSpecName: "scripts") pod "78919be1-5911-43e4-aa83-b8b06c443f4f" (UID: "78919be1-5911-43e4-aa83-b8b06c443f4f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.342223 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "78919be1-5911-43e4-aa83-b8b06c443f4f" (UID: "78919be1-5911-43e4-aa83-b8b06c443f4f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.343369 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78919be1-5911-43e4-aa83-b8b06c443f4f-kube-api-access-c4dbl" (OuterVolumeSpecName: "kube-api-access-c4dbl") pod "78919be1-5911-43e4-aa83-b8b06c443f4f" (UID: "78919be1-5911-43e4-aa83-b8b06c443f4f"). InnerVolumeSpecName "kube-api-access-c4dbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.358108 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78919be1-5911-43e4-aa83-b8b06c443f4f" (UID: "78919be1-5911-43e4-aa83-b8b06c443f4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.378937 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-config-data" (OuterVolumeSpecName: "config-data") pod "78919be1-5911-43e4-aa83-b8b06c443f4f" (UID: "78919be1-5911-43e4-aa83-b8b06c443f4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.438691 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.438724 4943 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.438735 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.438744 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4dbl\" (UniqueName: \"kubernetes.io/projected/78919be1-5911-43e4-aa83-b8b06c443f4f-kube-api-access-c4dbl\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.438754 4943 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78919be1-5911-43e4-aa83-b8b06c443f4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.924274 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-6kxff" event={"ID":"78919be1-5911-43e4-aa83-b8b06c443f4f","Type":"ContainerDied","Data":"31e0aeda0a452bf1a43c7de3f14fd54f6b9408104301ded7b771fa4998729cd9"} Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.924386 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31e0aeda0a452bf1a43c7de3f14fd54f6b9408104301ded7b771fa4998729cd9" Dec 04 10:37:42 crc kubenswrapper[4943]: I1204 10:37:42.924447 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-6kxff" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.424427 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 04 10:37:43 crc kubenswrapper[4943]: E1204 10:37:43.424707 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78919be1-5911-43e4-aa83-b8b06c443f4f" containerName="cinder-db-sync" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.424723 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="78919be1-5911-43e4-aa83-b8b06c443f4f" containerName="cinder-db-sync" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.424895 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="78919be1-5911-43e4-aa83-b8b06c443f4f" containerName="cinder-db-sync" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.425630 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.429917 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.430218 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"combined-ca-bundle" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.430348 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scheduler-config-data" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.430906 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-kf298" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.431254 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.452254 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.453687 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.455378 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-backup-config-data" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.470641 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.481840 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.482981 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.485685 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-volume-volume1-config-data" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.494771 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.527225 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.554654 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.555712 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.555990 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-dev\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556050 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556084 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556105 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556162 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556264 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556332 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556376 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556402 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-run\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556476 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-scripts\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556497 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556517 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgfbd\" (UniqueName: \"kubernetes.io/projected/363449dd-661a-484d-8e0d-a5b6e2a01f1a-kube-api-access-hgfbd\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556568 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmdvh\" (UniqueName: \"kubernetes.io/projected/37b1200a-9080-4647-bb43-844e4cd27f6d-kube-api-access-dmdvh\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556595 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556645 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556672 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556696 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556749 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-scripts\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556774 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/363449dd-661a-484d-8e0d-a5b6e2a01f1a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556801 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556838 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556915 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.556940 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557002 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557033 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-run\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557047 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-sys\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557064 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-sys\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557080 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557102 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-lib-modules\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557140 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7cks\" (UniqueName: \"kubernetes.io/projected/0dfd1962-064e-4544-b7b1-1b0ebed03aca-kube-api-access-m7cks\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557166 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-dev\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557180 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557260 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-config-data\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557354 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557400 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-config-data\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.557424 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.559144 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cert-cinder-public-svc" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.559219 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cert-cinder-internal-svc" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.559299 4943 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.563472 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.658782 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.658822 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.658844 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-run\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.658868 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.658898 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgfbd\" (UniqueName: \"kubernetes.io/projected/363449dd-661a-484d-8e0d-a5b6e2a01f1a-kube-api-access-hgfbd\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.658905 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-run\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.658925 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-scripts\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.658977 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.658896 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659006 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmdvh\" (UniqueName: \"kubernetes.io/projected/37b1200a-9080-4647-bb43-844e4cd27f6d-kube-api-access-dmdvh\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659067 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659103 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659131 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659158 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659178 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-config-data\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659215 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659241 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-scripts\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659256 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/363449dd-661a-484d-8e0d-a5b6e2a01f1a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659271 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659269 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659334 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659377 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659392 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/363449dd-661a-484d-8e0d-a5b6e2a01f1a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659447 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659287 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659488 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659503 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659494 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659530 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659550 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659573 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659623 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659642 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659657 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-run\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659673 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-sys\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659690 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-sys\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659693 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-run\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659705 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659724 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-lib-modules\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659746 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-config-data-custom\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659770 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7cks\" (UniqueName: \"kubernetes.io/projected/0dfd1962-064e-4544-b7b1-1b0ebed03aca-kube-api-access-m7cks\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659789 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-dev\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659805 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659828 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-config-data\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659845 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-logs\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659863 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659882 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-config-data\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659901 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659923 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-dev\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659938 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659954 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659971 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-scripts\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659987 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.660003 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.660017 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.660046 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xbgf\" (UniqueName: \"kubernetes.io/projected/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-kube-api-access-6xbgf\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.660070 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.660181 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.660244 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-lib-modules\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.659726 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-sys\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.660475 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-dev\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.662148 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-sys\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.662687 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.662707 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-dev\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.662772 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.663191 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-scripts\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.663283 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.663334 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.663695 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.663864 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.669409 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.669557 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-scripts\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.670979 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-config-data\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.671322 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.675122 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.677432 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-config-data\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.677529 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.679921 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.679960 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgfbd\" (UniqueName: \"kubernetes.io/projected/363449dd-661a-484d-8e0d-a5b6e2a01f1a-kube-api-access-hgfbd\") pod \"cinder-scheduler-0\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.680423 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7cks\" (UniqueName: \"kubernetes.io/projected/0dfd1962-064e-4544-b7b1-1b0ebed03aca-kube-api-access-m7cks\") pod \"cinder-backup-0\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.681080 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmdvh\" (UniqueName: \"kubernetes.io/projected/37b1200a-9080-4647-bb43-844e4cd27f6d-kube-api-access-dmdvh\") pod \"cinder-volume-volume1-0\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.742644 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.761512 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.761583 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-config-data-custom\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.761622 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-logs\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.761648 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.761666 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-scripts\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.761681 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xbgf\" (UniqueName: \"kubernetes.io/projected/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-kube-api-access-6xbgf\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.761715 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.761737 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.761755 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-config-data\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.762613 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-etc-machine-id\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.762924 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-logs\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.765050 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-public-tls-certs\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.765834 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-config-data\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.766265 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-config-data-custom\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.769012 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.769348 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-scripts\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.771239 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.773123 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.785785 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xbgf\" (UniqueName: \"kubernetes.io/projected/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-kube-api-access-6xbgf\") pod \"cinder-api-0\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.800889 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:43 crc kubenswrapper[4943]: I1204 10:37:43.871900 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:44 crc kubenswrapper[4943]: W1204 10:37:44.198164 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod363449dd_661a_484d_8e0d_a5b6e2a01f1a.slice/crio-f3097fd94c073764e8fac4aa0586e0f34b8bba18e8a027a185cae1bf68a78c91 WatchSource:0}: Error finding container f3097fd94c073764e8fac4aa0586e0f34b8bba18e8a027a185cae1bf68a78c91: Status 404 returned error can't find the container with id f3097fd94c073764e8fac4aa0586e0f34b8bba18e8a027a185cae1bf68a78c91 Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.199291 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.212350 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.282956 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 04 10:37:44 crc kubenswrapper[4943]: W1204 10:37:44.287786 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dfd1962_064e_4544_b7b1_1b0ebed03aca.slice/crio-266dfc843b7522d1f827f73879225aaa395b4058cfe8aeeec706ee3a77d9e8a2 WatchSource:0}: Error finding container 266dfc843b7522d1f827f73879225aaa395b4058cfe8aeeec706ee3a77d9e8a2: Status 404 returned error can't find the container with id 266dfc843b7522d1f827f73879225aaa395b4058cfe8aeeec706ee3a77d9e8a2 Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.369508 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.960855 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"37b1200a-9080-4647-bb43-844e4cd27f6d","Type":"ContainerStarted","Data":"90c830bd2bb83f6c6a3f16a9c0662aacb779af38f7c6409f93f528fd04854d8d"} Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.961463 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"37b1200a-9080-4647-bb43-844e4cd27f6d","Type":"ContainerStarted","Data":"ffef40e67f67f0d298c61442c487748d1b200645a397c9f16c662399d161414b"} Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.961487 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"37b1200a-9080-4647-bb43-844e4cd27f6d","Type":"ContainerStarted","Data":"ed216a22e68a87b7a058b60d5d64c72c1346a9e70c04fd25bb911392394614bc"} Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.971928 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"0dfd1962-064e-4544-b7b1-1b0ebed03aca","Type":"ContainerStarted","Data":"92f21c692ef65e85efafa3247f9223edf609ed0dd25f9a5e84c81dd33ca4e195"} Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.971975 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"0dfd1962-064e-4544-b7b1-1b0ebed03aca","Type":"ContainerStarted","Data":"4daafeb5bb61bd4a2c1752e5e1d776291a82271f30fb9cabf24ed952b333e675"} Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.971990 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"0dfd1962-064e-4544-b7b1-1b0ebed03aca","Type":"ContainerStarted","Data":"266dfc843b7522d1f827f73879225aaa395b4058cfe8aeeec706ee3a77d9e8a2"} Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.975639 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c","Type":"ContainerStarted","Data":"d8a7f476a9652a19719199f93a56e1b2ce6eb8d576781875c81e55ca29b4599f"} Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.977294 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"363449dd-661a-484d-8e0d-a5b6e2a01f1a","Type":"ContainerStarted","Data":"72fd684f702b689213c77b93bc36f31c5db3644c370d6c2015afa8b97bada4e8"} Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.977327 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"363449dd-661a-484d-8e0d-a5b6e2a01f1a","Type":"ContainerStarted","Data":"f3097fd94c073764e8fac4aa0586e0f34b8bba18e8a027a185cae1bf68a78c91"} Dec 04 10:37:44 crc kubenswrapper[4943]: I1204 10:37:44.988797 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podStartSLOduration=1.9877550739999998 podStartE2EDuration="1.987755074s" podCreationTimestamp="2025-12-04 10:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:37:44.985562113 +0000 UTC m=+1313.574537981" watchObservedRunningTime="2025-12-04 10:37:44.987755074 +0000 UTC m=+1313.576730962" Dec 04 10:37:45 crc kubenswrapper[4943]: I1204 10:37:45.987510 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c","Type":"ContainerStarted","Data":"b142c80685eecb0c61d1176610f4f7807cde73f6d5fbfc791813b3826d267b35"} Dec 04 10:37:45 crc kubenswrapper[4943]: I1204 10:37:45.988011 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:45 crc kubenswrapper[4943]: I1204 10:37:45.988022 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c","Type":"ContainerStarted","Data":"8e55d59849350e76d298bd622021d73f398a7bf749926b0fecac77b4c214cad1"} Dec 04 10:37:45 crc kubenswrapper[4943]: I1204 10:37:45.989684 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"363449dd-661a-484d-8e0d-a5b6e2a01f1a","Type":"ContainerStarted","Data":"416ab50df8cfad3e575cb13d1d5a4187d5a9aab8578931a2610712c81bc35fd9"} Dec 04 10:37:46 crc kubenswrapper[4943]: I1204 10:37:46.008395 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=3.008377044 podStartE2EDuration="3.008377044s" podCreationTimestamp="2025-12-04 10:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:37:46.006014719 +0000 UTC m=+1314.594990587" watchObservedRunningTime="2025-12-04 10:37:46.008377044 +0000 UTC m=+1314.597352912" Dec 04 10:37:46 crc kubenswrapper[4943]: I1204 10:37:46.009763 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-0" podStartSLOduration=3.009755382 podStartE2EDuration="3.009755382s" podCreationTimestamp="2025-12-04 10:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:37:45.010574352 +0000 UTC m=+1313.599550220" watchObservedRunningTime="2025-12-04 10:37:46.009755382 +0000 UTC m=+1314.598731250" Dec 04 10:37:46 crc kubenswrapper[4943]: I1204 10:37:46.038314 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-0" podStartSLOduration=3.038289618 podStartE2EDuration="3.038289618s" podCreationTimestamp="2025-12-04 10:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:37:46.029260519 +0000 UTC m=+1314.618236397" watchObservedRunningTime="2025-12-04 10:37:46.038289618 +0000 UTC m=+1314.627265486" Dec 04 10:37:46 crc kubenswrapper[4943]: I1204 10:37:46.998839 4943 generic.go:334] "Generic (PLEG): container finished" podID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerID="90c830bd2bb83f6c6a3f16a9c0662aacb779af38f7c6409f93f528fd04854d8d" exitCode=1 Dec 04 10:37:46 crc kubenswrapper[4943]: I1204 10:37:46.998934 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"37b1200a-9080-4647-bb43-844e4cd27f6d","Type":"ContainerDied","Data":"90c830bd2bb83f6c6a3f16a9c0662aacb779af38f7c6409f93f528fd04854d8d"} Dec 04 10:37:47 crc kubenswrapper[4943]: I1204 10:37:47.000132 4943 scope.go:117] "RemoveContainer" containerID="90c830bd2bb83f6c6a3f16a9c0662aacb779af38f7c6409f93f528fd04854d8d" Dec 04 10:37:48 crc kubenswrapper[4943]: I1204 10:37:48.008738 4943 generic.go:334] "Generic (PLEG): container finished" podID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerID="ffef40e67f67f0d298c61442c487748d1b200645a397c9f16c662399d161414b" exitCode=1 Dec 04 10:37:48 crc kubenswrapper[4943]: I1204 10:37:48.008801 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"37b1200a-9080-4647-bb43-844e4cd27f6d","Type":"ContainerDied","Data":"ffef40e67f67f0d298c61442c487748d1b200645a397c9f16c662399d161414b"} Dec 04 10:37:48 crc kubenswrapper[4943]: I1204 10:37:48.009075 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"37b1200a-9080-4647-bb43-844e4cd27f6d","Type":"ContainerStarted","Data":"3352778b6a7bd2c9136e4fed528aab88ae1c84125a331d66b74b2112409deb36"} Dec 04 10:37:48 crc kubenswrapper[4943]: I1204 10:37:48.009665 4943 scope.go:117] "RemoveContainer" containerID="ffef40e67f67f0d298c61442c487748d1b200645a397c9f16c662399d161414b" Dec 04 10:37:48 crc kubenswrapper[4943]: I1204 10:37:48.629456 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:37:48 crc kubenswrapper[4943]: I1204 10:37:48.629767 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:37:48 crc kubenswrapper[4943]: I1204 10:37:48.743503 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:48 crc kubenswrapper[4943]: I1204 10:37:48.771606 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:48 crc kubenswrapper[4943]: I1204 10:37:48.801453 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:48 crc kubenswrapper[4943]: I1204 10:37:48.801512 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:49 crc kubenswrapper[4943]: I1204 10:37:49.022134 4943 generic.go:334] "Generic (PLEG): container finished" podID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerID="3352778b6a7bd2c9136e4fed528aab88ae1c84125a331d66b74b2112409deb36" exitCode=1 Dec 04 10:37:49 crc kubenswrapper[4943]: I1204 10:37:49.023391 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"37b1200a-9080-4647-bb43-844e4cd27f6d","Type":"ContainerDied","Data":"3352778b6a7bd2c9136e4fed528aab88ae1c84125a331d66b74b2112409deb36"} Dec 04 10:37:49 crc kubenswrapper[4943]: I1204 10:37:49.023490 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"37b1200a-9080-4647-bb43-844e4cd27f6d","Type":"ContainerStarted","Data":"294bb719ff8e73c5d3d235f47484d4f98792d4febfed0703810a0203cf5b3c95"} Dec 04 10:37:49 crc kubenswrapper[4943]: I1204 10:37:49.023523 4943 scope.go:117] "RemoveContainer" containerID="90c830bd2bb83f6c6a3f16a9c0662aacb779af38f7c6409f93f528fd04854d8d" Dec 04 10:37:49 crc kubenswrapper[4943]: I1204 10:37:49.024284 4943 scope.go:117] "RemoveContainer" containerID="3352778b6a7bd2c9136e4fed528aab88ae1c84125a331d66b74b2112409deb36" Dec 04 10:37:49 crc kubenswrapper[4943]: E1204 10:37:49.024889 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(37b1200a-9080-4647-bb43-844e4cd27f6d)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" Dec 04 10:37:50 crc kubenswrapper[4943]: I1204 10:37:50.030809 4943 generic.go:334] "Generic (PLEG): container finished" podID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerID="294bb719ff8e73c5d3d235f47484d4f98792d4febfed0703810a0203cf5b3c95" exitCode=1 Dec 04 10:37:50 crc kubenswrapper[4943]: I1204 10:37:50.030929 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"37b1200a-9080-4647-bb43-844e4cd27f6d","Type":"ContainerDied","Data":"294bb719ff8e73c5d3d235f47484d4f98792d4febfed0703810a0203cf5b3c95"} Dec 04 10:37:50 crc kubenswrapper[4943]: I1204 10:37:50.031165 4943 scope.go:117] "RemoveContainer" containerID="ffef40e67f67f0d298c61442c487748d1b200645a397c9f16c662399d161414b" Dec 04 10:37:50 crc kubenswrapper[4943]: I1204 10:37:50.031707 4943 scope.go:117] "RemoveContainer" containerID="294bb719ff8e73c5d3d235f47484d4f98792d4febfed0703810a0203cf5b3c95" Dec 04 10:37:50 crc kubenswrapper[4943]: I1204 10:37:50.031736 4943 scope.go:117] "RemoveContainer" containerID="3352778b6a7bd2c9136e4fed528aab88ae1c84125a331d66b74b2112409deb36" Dec 04 10:37:50 crc kubenswrapper[4943]: E1204 10:37:50.033123 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(37b1200a-9080-4647-bb43-844e4cd27f6d)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(37b1200a-9080-4647-bb43-844e4cd27f6d)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" Dec 04 10:37:51 crc kubenswrapper[4943]: I1204 10:37:51.042866 4943 scope.go:117] "RemoveContainer" containerID="294bb719ff8e73c5d3d235f47484d4f98792d4febfed0703810a0203cf5b3c95" Dec 04 10:37:51 crc kubenswrapper[4943]: I1204 10:37:51.042896 4943 scope.go:117] "RemoveContainer" containerID="3352778b6a7bd2c9136e4fed528aab88ae1c84125a331d66b74b2112409deb36" Dec 04 10:37:51 crc kubenswrapper[4943]: E1204 10:37:51.043121 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(37b1200a-9080-4647-bb43-844e4cd27f6d)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(37b1200a-9080-4647-bb43-844e4cd27f6d)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" Dec 04 10:37:52 crc kubenswrapper[4943]: I1204 10:37:52.801334 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:52 crc kubenswrapper[4943]: I1204 10:37:52.802842 4943 scope.go:117] "RemoveContainer" containerID="294bb719ff8e73c5d3d235f47484d4f98792d4febfed0703810a0203cf5b3c95" Dec 04 10:37:52 crc kubenswrapper[4943]: I1204 10:37:52.802860 4943 scope.go:117] "RemoveContainer" containerID="3352778b6a7bd2c9136e4fed528aab88ae1c84125a331d66b74b2112409deb36" Dec 04 10:37:52 crc kubenswrapper[4943]: E1204 10:37:52.803145 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(37b1200a-9080-4647-bb43-844e4cd27f6d)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(37b1200a-9080-4647-bb43-844e4cd27f6d)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" Dec 04 10:37:53 crc kubenswrapper[4943]: I1204 10:37:53.801335 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:53 crc kubenswrapper[4943]: I1204 10:37:53.801411 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:53 crc kubenswrapper[4943]: I1204 10:37:53.802348 4943 scope.go:117] "RemoveContainer" containerID="294bb719ff8e73c5d3d235f47484d4f98792d4febfed0703810a0203cf5b3c95" Dec 04 10:37:53 crc kubenswrapper[4943]: I1204 10:37:53.802395 4943 scope.go:117] "RemoveContainer" containerID="3352778b6a7bd2c9136e4fed528aab88ae1c84125a331d66b74b2112409deb36" Dec 04 10:37:53 crc kubenswrapper[4943]: E1204 10:37:53.802803 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(37b1200a-9080-4647-bb43-844e4cd27f6d)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(37b1200a-9080-4647-bb43-844e4cd27f6d)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" Dec 04 10:37:53 crc kubenswrapper[4943]: I1204 10:37:53.941698 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:37:54 crc kubenswrapper[4943]: I1204 10:37:54.000346 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:37:55 crc kubenswrapper[4943]: I1204 10:37:55.754482 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.344919 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-6kxff"] Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.352333 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-6kxff"] Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.364614 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.364906 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="0dfd1962-064e-4544-b7b1-1b0ebed03aca" containerName="cinder-backup" containerID="cri-o://4daafeb5bb61bd4a2c1752e5e1d776291a82271f30fb9cabf24ed952b333e675" gracePeriod=30 Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.364994 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="0dfd1962-064e-4544-b7b1-1b0ebed03aca" containerName="probe" containerID="cri-o://92f21c692ef65e85efafa3247f9223edf609ed0dd25f9a5e84c81dd33ca4e195" gracePeriod=30 Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.409419 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.409683 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" containerName="cinder-api-log" containerID="cri-o://8e55d59849350e76d298bd622021d73f398a7bf749926b0fecac77b4c214cad1" gracePeriod=30 Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.410130 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" containerName="cinder-api" containerID="cri-o://b142c80685eecb0c61d1176610f4f7807cde73f6d5fbfc791813b3826d267b35" gracePeriod=30 Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.417627 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="cinder-kuttl-tests/cinder-api-0" podUID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.107:8776/healthcheck\": EOF" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.419209 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.430303 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.430644 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="363449dd-661a-484d-8e0d-a5b6e2a01f1a" containerName="cinder-scheduler" containerID="cri-o://72fd684f702b689213c77b93bc36f31c5db3644c370d6c2015afa8b97bada4e8" gracePeriod=30 Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.430825 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="363449dd-661a-484d-8e0d-a5b6e2a01f1a" containerName="probe" containerID="cri-o://416ab50df8cfad3e575cb13d1d5a4187d5a9aab8578931a2610712c81bc35fd9" gracePeriod=30 Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.438058 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder6acd-account-delete-mg8sx"] Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.446017 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.451546 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder6acd-account-delete-mg8sx"] Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.602002 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04760d52-535c-47b9-81ee-a0d5e0e6b965-operator-scripts\") pod \"cinder6acd-account-delete-mg8sx\" (UID: \"04760d52-535c-47b9-81ee-a0d5e0e6b965\") " pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.602074 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4c9w\" (UniqueName: \"kubernetes.io/projected/04760d52-535c-47b9-81ee-a0d5e0e6b965-kube-api-access-z4c9w\") pod \"cinder6acd-account-delete-mg8sx\" (UID: \"04760d52-535c-47b9-81ee-a0d5e0e6b965\") " pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.703535 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4c9w\" (UniqueName: \"kubernetes.io/projected/04760d52-535c-47b9-81ee-a0d5e0e6b965-kube-api-access-z4c9w\") pod \"cinder6acd-account-delete-mg8sx\" (UID: \"04760d52-535c-47b9-81ee-a0d5e0e6b965\") " pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.703691 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04760d52-535c-47b9-81ee-a0d5e0e6b965-operator-scripts\") pod \"cinder6acd-account-delete-mg8sx\" (UID: \"04760d52-535c-47b9-81ee-a0d5e0e6b965\") " pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.704474 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04760d52-535c-47b9-81ee-a0d5e0e6b965-operator-scripts\") pod \"cinder6acd-account-delete-mg8sx\" (UID: \"04760d52-535c-47b9-81ee-a0d5e0e6b965\") " pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.724871 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4c9w\" (UniqueName: \"kubernetes.io/projected/04760d52-535c-47b9-81ee-a0d5e0e6b965-kube-api-access-z4c9w\") pod \"cinder6acd-account-delete-mg8sx\" (UID: \"04760d52-535c-47b9-81ee-a0d5e0e6b965\") " pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.766673 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.769960 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.906774 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-dev\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.906866 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-locks-brick\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.906889 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-combined-ca-bundle\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.906944 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-nvme\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.906973 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-lib-cinder\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.906994 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-lib-modules\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907013 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-sys\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907032 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-config-data\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907049 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-run\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907067 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-iscsi\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907107 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-scripts\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907136 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-machine-id\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907173 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-locks-cinder\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907225 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-config-data-custom\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907280 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmdvh\" (UniqueName: \"kubernetes.io/projected/37b1200a-9080-4647-bb43-844e4cd27f6d-kube-api-access-dmdvh\") pod \"37b1200a-9080-4647-bb43-844e4cd27f6d\" (UID: \"37b1200a-9080-4647-bb43-844e4cd27f6d\") " Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907583 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-sys" (OuterVolumeSpecName: "sys") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907643 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-dev" (OuterVolumeSpecName: "dev") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907668 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907786 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907851 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.907874 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-run" (OuterVolumeSpecName: "run") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.908279 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.908541 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.908560 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.908579 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.909996 4943 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.910019 4943 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-dev\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.910031 4943 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.910043 4943 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.910054 4943 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.910067 4943 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.910078 4943 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-sys\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.910088 4943 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-run\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.910097 4943 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.910108 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/37b1200a-9080-4647-bb43-844e4cd27f6d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.926485 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.926573 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-scripts" (OuterVolumeSpecName: "scripts") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.926687 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37b1200a-9080-4647-bb43-844e4cd27f6d-kube-api-access-dmdvh" (OuterVolumeSpecName: "kube-api-access-dmdvh") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "kube-api-access-dmdvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.954651 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:57 crc kubenswrapper[4943]: I1204 10:37:57.986469 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-config-data" (OuterVolumeSpecName: "config-data") pod "37b1200a-9080-4647-bb43-844e4cd27f6d" (UID: "37b1200a-9080-4647-bb43-844e4cd27f6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.011591 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.011624 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.011635 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.011645 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmdvh\" (UniqueName: \"kubernetes.io/projected/37b1200a-9080-4647-bb43-844e4cd27f6d-kube-api-access-dmdvh\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.011654 4943 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b1200a-9080-4647-bb43-844e4cd27f6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.099371 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"37b1200a-9080-4647-bb43-844e4cd27f6d","Type":"ContainerDied","Data":"ed216a22e68a87b7a058b60d5d64c72c1346a9e70c04fd25bb911392394614bc"} Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.099420 4943 scope.go:117] "RemoveContainer" containerID="294bb719ff8e73c5d3d235f47484d4f98792d4febfed0703810a0203cf5b3c95" Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.099512 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.109456 4943 generic.go:334] "Generic (PLEG): container finished" podID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" containerID="8e55d59849350e76d298bd622021d73f398a7bf749926b0fecac77b4c214cad1" exitCode=143 Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.109510 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c","Type":"ContainerDied","Data":"8e55d59849350e76d298bd622021d73f398a7bf749926b0fecac77b4c214cad1"} Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.135678 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.141836 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.149856 4943 scope.go:117] "RemoveContainer" containerID="3352778b6a7bd2c9136e4fed528aab88ae1c84125a331d66b74b2112409deb36" Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.190066 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder6acd-account-delete-mg8sx"] Dec 04 10:37:58 crc kubenswrapper[4943]: W1204 10:37:58.199316 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04760d52_535c_47b9_81ee_a0d5e0e6b965.slice/crio-c7697c5771f7a50fdaa061d1ee1bf7d7f61a86e9b5e08bcfaa3bb5a51d0f2dbb WatchSource:0}: Error finding container c7697c5771f7a50fdaa061d1ee1bf7d7f61a86e9b5e08bcfaa3bb5a51d0f2dbb: Status 404 returned error can't find the container with id c7697c5771f7a50fdaa061d1ee1bf7d7f61a86e9b5e08bcfaa3bb5a51d0f2dbb Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.574087 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" path="/var/lib/kubelet/pods/37b1200a-9080-4647-bb43-844e4cd27f6d/volumes" Dec 04 10:37:58 crc kubenswrapper[4943]: I1204 10:37:58.575288 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78919be1-5911-43e4-aa83-b8b06c443f4f" path="/var/lib/kubelet/pods/78919be1-5911-43e4-aa83-b8b06c443f4f/volumes" Dec 04 10:37:59 crc kubenswrapper[4943]: I1204 10:37:59.118641 4943 generic.go:334] "Generic (PLEG): container finished" podID="0dfd1962-064e-4544-b7b1-1b0ebed03aca" containerID="92f21c692ef65e85efafa3247f9223edf609ed0dd25f9a5e84c81dd33ca4e195" exitCode=0 Dec 04 10:37:59 crc kubenswrapper[4943]: I1204 10:37:59.118730 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"0dfd1962-064e-4544-b7b1-1b0ebed03aca","Type":"ContainerDied","Data":"92f21c692ef65e85efafa3247f9223edf609ed0dd25f9a5e84c81dd33ca4e195"} Dec 04 10:37:59 crc kubenswrapper[4943]: I1204 10:37:59.120998 4943 generic.go:334] "Generic (PLEG): container finished" podID="363449dd-661a-484d-8e0d-a5b6e2a01f1a" containerID="416ab50df8cfad3e575cb13d1d5a4187d5a9aab8578931a2610712c81bc35fd9" exitCode=0 Dec 04 10:37:59 crc kubenswrapper[4943]: I1204 10:37:59.121101 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"363449dd-661a-484d-8e0d-a5b6e2a01f1a","Type":"ContainerDied","Data":"416ab50df8cfad3e575cb13d1d5a4187d5a9aab8578931a2610712c81bc35fd9"} Dec 04 10:37:59 crc kubenswrapper[4943]: I1204 10:37:59.122818 4943 generic.go:334] "Generic (PLEG): container finished" podID="04760d52-535c-47b9-81ee-a0d5e0e6b965" containerID="2851d6f7b6ae0d36fedbeb621349219674f4e6fcd07792586309ddb4a6a1563b" exitCode=0 Dec 04 10:37:59 crc kubenswrapper[4943]: I1204 10:37:59.122879 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" event={"ID":"04760d52-535c-47b9-81ee-a0d5e0e6b965","Type":"ContainerDied","Data":"2851d6f7b6ae0d36fedbeb621349219674f4e6fcd07792586309ddb4a6a1563b"} Dec 04 10:37:59 crc kubenswrapper[4943]: I1204 10:37:59.122898 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" event={"ID":"04760d52-535c-47b9-81ee-a0d5e0e6b965","Type":"ContainerStarted","Data":"c7697c5771f7a50fdaa061d1ee1bf7d7f61a86e9b5e08bcfaa3bb5a51d0f2dbb"} Dec 04 10:38:00 crc kubenswrapper[4943]: I1204 10:38:00.385004 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" Dec 04 10:38:00 crc kubenswrapper[4943]: I1204 10:38:00.550140 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04760d52-535c-47b9-81ee-a0d5e0e6b965-operator-scripts\") pod \"04760d52-535c-47b9-81ee-a0d5e0e6b965\" (UID: \"04760d52-535c-47b9-81ee-a0d5e0e6b965\") " Dec 04 10:38:00 crc kubenswrapper[4943]: I1204 10:38:00.550322 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4c9w\" (UniqueName: \"kubernetes.io/projected/04760d52-535c-47b9-81ee-a0d5e0e6b965-kube-api-access-z4c9w\") pod \"04760d52-535c-47b9-81ee-a0d5e0e6b965\" (UID: \"04760d52-535c-47b9-81ee-a0d5e0e6b965\") " Dec 04 10:38:00 crc kubenswrapper[4943]: I1204 10:38:00.550926 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04760d52-535c-47b9-81ee-a0d5e0e6b965-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "04760d52-535c-47b9-81ee-a0d5e0e6b965" (UID: "04760d52-535c-47b9-81ee-a0d5e0e6b965"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:00 crc kubenswrapper[4943]: I1204 10:38:00.555769 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04760d52-535c-47b9-81ee-a0d5e0e6b965-kube-api-access-z4c9w" (OuterVolumeSpecName: "kube-api-access-z4c9w") pod "04760d52-535c-47b9-81ee-a0d5e0e6b965" (UID: "04760d52-535c-47b9-81ee-a0d5e0e6b965"). InnerVolumeSpecName "kube-api-access-z4c9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:00 crc kubenswrapper[4943]: I1204 10:38:00.653704 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04760d52-535c-47b9-81ee-a0d5e0e6b965-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:00 crc kubenswrapper[4943]: I1204 10:38:00.653768 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4c9w\" (UniqueName: \"kubernetes.io/projected/04760d52-535c-47b9-81ee-a0d5e0e6b965-kube-api-access-z4c9w\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.141796 4943 generic.go:334] "Generic (PLEG): container finished" podID="0dfd1962-064e-4544-b7b1-1b0ebed03aca" containerID="4daafeb5bb61bd4a2c1752e5e1d776291a82271f30fb9cabf24ed952b333e675" exitCode=0 Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.141883 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"0dfd1962-064e-4544-b7b1-1b0ebed03aca","Type":"ContainerDied","Data":"4daafeb5bb61bd4a2c1752e5e1d776291a82271f30fb9cabf24ed952b333e675"} Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.144877 4943 generic.go:334] "Generic (PLEG): container finished" podID="363449dd-661a-484d-8e0d-a5b6e2a01f1a" containerID="72fd684f702b689213c77b93bc36f31c5db3644c370d6c2015afa8b97bada4e8" exitCode=0 Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.144916 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"363449dd-661a-484d-8e0d-a5b6e2a01f1a","Type":"ContainerDied","Data":"72fd684f702b689213c77b93bc36f31c5db3644c370d6c2015afa8b97bada4e8"} Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.147307 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" event={"ID":"04760d52-535c-47b9-81ee-a0d5e0e6b965","Type":"ContainerDied","Data":"c7697c5771f7a50fdaa061d1ee1bf7d7f61a86e9b5e08bcfaa3bb5a51d0f2dbb"} Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.147369 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7697c5771f7a50fdaa061d1ee1bf7d7f61a86e9b5e08bcfaa3bb5a51d0f2dbb" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.147412 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder6acd-account-delete-mg8sx" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.255977 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.279407 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-config-data-custom\") pod \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.279470 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-combined-ca-bundle\") pod \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.279486 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-config-data\") pod \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.279521 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.279530 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgfbd\" (UniqueName: \"kubernetes.io/projected/363449dd-661a-484d-8e0d-a5b6e2a01f1a-kube-api-access-hgfbd\") pod \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.279632 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-scripts\") pod \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.286973 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "363449dd-661a-484d-8e0d-a5b6e2a01f1a" (UID: "363449dd-661a-484d-8e0d-a5b6e2a01f1a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.286992 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-scripts" (OuterVolumeSpecName: "scripts") pod "363449dd-661a-484d-8e0d-a5b6e2a01f1a" (UID: "363449dd-661a-484d-8e0d-a5b6e2a01f1a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.291065 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/363449dd-661a-484d-8e0d-a5b6e2a01f1a-kube-api-access-hgfbd" (OuterVolumeSpecName: "kube-api-access-hgfbd") pod "363449dd-661a-484d-8e0d-a5b6e2a01f1a" (UID: "363449dd-661a-484d-8e0d-a5b6e2a01f1a"). InnerVolumeSpecName "kube-api-access-hgfbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.343027 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-config-data" (OuterVolumeSpecName: "config-data") pod "363449dd-661a-484d-8e0d-a5b6e2a01f1a" (UID: "363449dd-661a-484d-8e0d-a5b6e2a01f1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.343686 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "363449dd-661a-484d-8e0d-a5b6e2a01f1a" (UID: "363449dd-661a-484d-8e0d-a5b6e2a01f1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381184 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7cks\" (UniqueName: \"kubernetes.io/projected/0dfd1962-064e-4544-b7b1-1b0ebed03aca-kube-api-access-m7cks\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381271 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-run\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381325 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-config-data-custom\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381351 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-nvme\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381364 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-dev\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381394 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-iscsi\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381408 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-sys\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381443 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-lib-cinder\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381462 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-machine-id\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381490 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-scripts\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381512 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-locks-cinder\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381542 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-lib-modules\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381574 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-config-data\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381599 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/363449dd-661a-484d-8e0d-a5b6e2a01f1a-etc-machine-id\") pod \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\" (UID: \"363449dd-661a-484d-8e0d-a5b6e2a01f1a\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381630 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-locks-brick\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381665 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-combined-ca-bundle\") pod \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\" (UID: \"0dfd1962-064e-4544-b7b1-1b0ebed03aca\") " Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381660 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381708 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-run" (OuterVolumeSpecName: "run") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381722 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381742 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381747 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-sys" (OuterVolumeSpecName: "sys") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381762 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-dev" (OuterVolumeSpecName: "dev") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381769 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381781 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381814 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/363449dd-661a-484d-8e0d-a5b6e2a01f1a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "363449dd-661a-484d-8e0d-a5b6e2a01f1a" (UID: "363449dd-661a-484d-8e0d-a5b6e2a01f1a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.381861 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382029 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382050 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382062 4943 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382074 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgfbd\" (UniqueName: \"kubernetes.io/projected/363449dd-661a-484d-8e0d-a5b6e2a01f1a-kube-api-access-hgfbd\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382088 4943 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-run\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382098 4943 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-nvme\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382108 4943 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-dev\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382118 4943 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-iscsi\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382168 4943 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-sys\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382178 4943 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382175 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382189 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382251 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/363449dd-661a-484d-8e0d-a5b6e2a01f1a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382263 4943 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382274 4943 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-lib-modules\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.382285 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/363449dd-661a-484d-8e0d-a5b6e2a01f1a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.384111 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dfd1962-064e-4544-b7b1-1b0ebed03aca-kube-api-access-m7cks" (OuterVolumeSpecName: "kube-api-access-m7cks") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "kube-api-access-m7cks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.384523 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-scripts" (OuterVolumeSpecName: "scripts") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.387428 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.418973 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.452752 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-config-data" (OuterVolumeSpecName: "config-data") pod "0dfd1962-064e-4544-b7b1-1b0ebed03aca" (UID: "0dfd1962-064e-4544-b7b1-1b0ebed03aca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.482939 4943 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0dfd1962-064e-4544-b7b1-1b0ebed03aca-var-locks-brick\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.482969 4943 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.482978 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7cks\" (UniqueName: \"kubernetes.io/projected/0dfd1962-064e-4544-b7b1-1b0ebed03aca-kube-api-access-m7cks\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.482991 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.483002 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.483009 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dfd1962-064e-4544-b7b1-1b0ebed03aca-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:01 crc kubenswrapper[4943]: I1204 10:38:01.828598 4943 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/cinder-api-0" podUID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.107:8776/healthcheck\": read tcp 10.217.0.2:44886->10.217.0.107:8776: read: connection reset by peer" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.160833 4943 generic.go:334] "Generic (PLEG): container finished" podID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" containerID="b142c80685eecb0c61d1176610f4f7807cde73f6d5fbfc791813b3826d267b35" exitCode=0 Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.160927 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c","Type":"ContainerDied","Data":"b142c80685eecb0c61d1176610f4f7807cde73f6d5fbfc791813b3826d267b35"} Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.163647 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"0dfd1962-064e-4544-b7b1-1b0ebed03aca","Type":"ContainerDied","Data":"266dfc843b7522d1f827f73879225aaa395b4058cfe8aeeec706ee3a77d9e8a2"} Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.163685 4943 scope.go:117] "RemoveContainer" containerID="92f21c692ef65e85efafa3247f9223edf609ed0dd25f9a5e84c81dd33ca4e195" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.163844 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.168475 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"363449dd-661a-484d-8e0d-a5b6e2a01f1a","Type":"ContainerDied","Data":"f3097fd94c073764e8fac4aa0586e0f34b8bba18e8a027a185cae1bf68a78c91"} Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.168587 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.188514 4943 scope.go:117] "RemoveContainer" containerID="4daafeb5bb61bd4a2c1752e5e1d776291a82271f30fb9cabf24ed952b333e675" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.205339 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.220159 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.220544 4943 scope.go:117] "RemoveContainer" containerID="416ab50df8cfad3e575cb13d1d5a4187d5a9aab8578931a2610712c81bc35fd9" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.221354 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.226587 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.231575 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.236736 4943 scope.go:117] "RemoveContainer" containerID="72fd684f702b689213c77b93bc36f31c5db3644c370d6c2015afa8b97bada4e8" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.397257 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-config-data-custom\") pod \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.397312 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-config-data\") pod \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.397397 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-scripts\") pod \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.397426 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-etc-machine-id\") pod \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.397444 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xbgf\" (UniqueName: \"kubernetes.io/projected/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-kube-api-access-6xbgf\") pod \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.397462 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-combined-ca-bundle\") pod \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.397492 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-logs\") pod \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.397533 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-internal-tls-certs\") pod \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.397614 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-public-tls-certs\") pod \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\" (UID: \"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c\") " Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.397651 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" (UID: "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.397895 4943 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.398527 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-logs" (OuterVolumeSpecName: "logs") pod "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" (UID: "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.403338 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-scripts" (OuterVolumeSpecName: "scripts") pod "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" (UID: "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.403388 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-kube-api-access-6xbgf" (OuterVolumeSpecName: "kube-api-access-6xbgf") pod "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" (UID: "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c"). InnerVolumeSpecName "kube-api-access-6xbgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.405488 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" (UID: "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.433571 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-d476h"] Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.434075 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" (UID: "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.451049 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-d476h"] Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.456091 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" (UID: "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.457813 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder6acd-account-delete-mg8sx"] Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.457863 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-config-data" (OuterVolumeSpecName: "config-data") pod "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" (UID: "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.464371 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" (UID: "75af066d-f8e0-4da5-9f9f-2a0bc9425a0c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.468519 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv"] Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.476153 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-6acd-account-create-update-ts5vv"] Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.480528 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder6acd-account-delete-mg8sx"] Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.499823 4943 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.499904 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.499947 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.499962 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.499975 4943 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.500039 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xbgf\" (UniqueName: \"kubernetes.io/projected/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-kube-api-access-6xbgf\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.500053 4943 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-logs\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.500065 4943 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.580756 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04760d52-535c-47b9-81ee-a0d5e0e6b965" path="/var/lib/kubelet/pods/04760d52-535c-47b9-81ee-a0d5e0e6b965/volumes" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.581599 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dfd1962-064e-4544-b7b1-1b0ebed03aca" path="/var/lib/kubelet/pods/0dfd1962-064e-4544-b7b1-1b0ebed03aca/volumes" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.582126 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1565f2bc-590f-4b50-ab58-05565b6df99c" path="/var/lib/kubelet/pods/1565f2bc-590f-4b50-ab58-05565b6df99c/volumes" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.583041 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="363449dd-661a-484d-8e0d-a5b6e2a01f1a" path="/var/lib/kubelet/pods/363449dd-661a-484d-8e0d-a5b6e2a01f1a/volumes" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:02.583575 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c80881c-bab1-4465-82f0-dcedc10fe88c" path="/var/lib/kubelet/pods/8c80881c-bab1-4465-82f0-dcedc10fe88c/volumes" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:03.179737 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:03.179728 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"75af066d-f8e0-4da5-9f9f-2a0bc9425a0c","Type":"ContainerDied","Data":"d8a7f476a9652a19719199f93a56e1b2ce6eb8d576781875c81e55ca29b4599f"} Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:03.179923 4943 scope.go:117] "RemoveContainer" containerID="b142c80685eecb0c61d1176610f4f7807cde73f6d5fbfc791813b3826d267b35" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:03.216738 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:03.217079 4943 scope.go:117] "RemoveContainer" containerID="8e55d59849350e76d298bd622021d73f398a7bf749926b0fecac77b4c214cad1" Dec 04 10:38:03 crc kubenswrapper[4943]: I1204 10:38:03.228031 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.251160 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-crh8b"] Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.257894 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-q94wl"] Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.269798 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-crh8b"] Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.275463 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-q94wl"] Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.281303 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2"] Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.281629 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" podUID="93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a" containerName="keystone-api" containerID="cri-o://5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf" gracePeriod=30 Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320191 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystonefea1-account-delete-c7hkh"] Dec 04 10:38:04 crc kubenswrapper[4943]: E1204 10:38:04.320520 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerName="cinder-volume" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320539 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerName="cinder-volume" Dec 04 10:38:04 crc kubenswrapper[4943]: E1204 10:38:04.320551 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363449dd-661a-484d-8e0d-a5b6e2a01f1a" containerName="probe" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320559 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="363449dd-661a-484d-8e0d-a5b6e2a01f1a" containerName="probe" Dec 04 10:38:04 crc kubenswrapper[4943]: E1204 10:38:04.320577 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dfd1962-064e-4544-b7b1-1b0ebed03aca" containerName="probe" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320586 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dfd1962-064e-4544-b7b1-1b0ebed03aca" containerName="probe" Dec 04 10:38:04 crc kubenswrapper[4943]: E1204 10:38:04.320600 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dfd1962-064e-4544-b7b1-1b0ebed03aca" containerName="cinder-backup" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320609 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dfd1962-064e-4544-b7b1-1b0ebed03aca" containerName="cinder-backup" Dec 04 10:38:04 crc kubenswrapper[4943]: E1204 10:38:04.320618 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04760d52-535c-47b9-81ee-a0d5e0e6b965" containerName="mariadb-account-delete" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320626 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="04760d52-535c-47b9-81ee-a0d5e0e6b965" containerName="mariadb-account-delete" Dec 04 10:38:04 crc kubenswrapper[4943]: E1204 10:38:04.320645 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363449dd-661a-484d-8e0d-a5b6e2a01f1a" containerName="cinder-scheduler" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320652 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="363449dd-661a-484d-8e0d-a5b6e2a01f1a" containerName="cinder-scheduler" Dec 04 10:38:04 crc kubenswrapper[4943]: E1204 10:38:04.320661 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" containerName="cinder-api-log" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320668 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" containerName="cinder-api-log" Dec 04 10:38:04 crc kubenswrapper[4943]: E1204 10:38:04.320682 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerName="probe" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320690 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerName="probe" Dec 04 10:38:04 crc kubenswrapper[4943]: E1204 10:38:04.320702 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" containerName="cinder-api" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320710 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" containerName="cinder-api" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320858 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dfd1962-064e-4544-b7b1-1b0ebed03aca" containerName="cinder-backup" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320873 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dfd1962-064e-4544-b7b1-1b0ebed03aca" containerName="probe" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320882 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerName="probe" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320894 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="363449dd-661a-484d-8e0d-a5b6e2a01f1a" containerName="probe" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320905 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerName="cinder-volume" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320914 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" containerName="cinder-api" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320923 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" containerName="cinder-api-log" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320933 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="363449dd-661a-484d-8e0d-a5b6e2a01f1a" containerName="cinder-scheduler" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.320948 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="04760d52-535c-47b9-81ee-a0d5e0e6b965" containerName="mariadb-account-delete" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.322133 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.330179 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystonefea1-account-delete-c7hkh"] Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.334230 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-operator-scripts\") pod \"keystonefea1-account-delete-c7hkh\" (UID: \"ba5c9e69-8d95-43ab-aff2-a2a7447c4920\") " pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.334290 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jnf7\" (UniqueName: \"kubernetes.io/projected/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-kube-api-access-6jnf7\") pod \"keystonefea1-account-delete-c7hkh\" (UID: \"ba5c9e69-8d95-43ab-aff2-a2a7447c4920\") " pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.435650 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-operator-scripts\") pod \"keystonefea1-account-delete-c7hkh\" (UID: \"ba5c9e69-8d95-43ab-aff2-a2a7447c4920\") " pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.435745 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jnf7\" (UniqueName: \"kubernetes.io/projected/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-kube-api-access-6jnf7\") pod \"keystonefea1-account-delete-c7hkh\" (UID: \"ba5c9e69-8d95-43ab-aff2-a2a7447c4920\") " pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.436713 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-operator-scripts\") pod \"keystonefea1-account-delete-c7hkh\" (UID: \"ba5c9e69-8d95-43ab-aff2-a2a7447c4920\") " pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.468846 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jnf7\" (UniqueName: \"kubernetes.io/projected/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-kube-api-access-6jnf7\") pod \"keystonefea1-account-delete-c7hkh\" (UID: \"ba5c9e69-8d95-43ab-aff2-a2a7447c4920\") " pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.573093 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06955606-470a-4395-a345-42d5170fa271" path="/var/lib/kubelet/pods/06955606-470a-4395-a345-42d5170fa271/volumes" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.573866 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75af066d-f8e0-4da5-9f9f-2a0bc9425a0c" path="/var/lib/kubelet/pods/75af066d-f8e0-4da5-9f9f-2a0bc9425a0c/volumes" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.574552 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaf4529b-f939-41eb-b31e-372ae95e1d51" path="/var/lib/kubelet/pods/eaf4529b-f939-41eb-b31e-372ae95e1d51/volumes" Dec 04 10:38:04 crc kubenswrapper[4943]: I1204 10:38:04.642139 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" Dec 04 10:38:05 crc kubenswrapper[4943]: I1204 10:38:05.052411 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystonefea1-account-delete-c7hkh"] Dec 04 10:38:05 crc kubenswrapper[4943]: I1204 10:38:05.109324 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 04 10:38:05 crc kubenswrapper[4943]: I1204 10:38:05.114571 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 04 10:38:05 crc kubenswrapper[4943]: I1204 10:38:05.123733 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 04 10:38:05 crc kubenswrapper[4943]: I1204 10:38:05.203417 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" event={"ID":"ba5c9e69-8d95-43ab-aff2-a2a7447c4920","Type":"ContainerStarted","Data":"743549dc69b994c64fff097be4a625dd0ddf654116af8581d4a64411ed140e7e"} Dec 04 10:38:05 crc kubenswrapper[4943]: I1204 10:38:05.272845 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/openstack-galera-2" podUID="90592895-1eb8-4b3b-bdb3-746b364256c1" containerName="galera" containerID="cri-o://f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286" gracePeriod=30 Dec 04 10:38:05 crc kubenswrapper[4943]: I1204 10:38:05.788298 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 04 10:38:05 crc kubenswrapper[4943]: I1204 10:38:05.788515 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/memcached-0" podUID="197663a6-bdeb-49ee-ad21-9b76212842e9" containerName="memcached" containerID="cri-o://982cf322919a66576035509dc13c52ef8d2f4017b05fe27ba6629d63ee3e4f1a" gracePeriod=30 Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.069430 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.160701 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-kolla-config\") pod \"90592895-1eb8-4b3b-bdb3-746b364256c1\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.160767 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/90592895-1eb8-4b3b-bdb3-746b364256c1-config-data-generated\") pod \"90592895-1eb8-4b3b-bdb3-746b364256c1\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.160797 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"90592895-1eb8-4b3b-bdb3-746b364256c1\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.160850 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdcfx\" (UniqueName: \"kubernetes.io/projected/90592895-1eb8-4b3b-bdb3-746b364256c1-kube-api-access-gdcfx\") pod \"90592895-1eb8-4b3b-bdb3-746b364256c1\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.160884 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-operator-scripts\") pod \"90592895-1eb8-4b3b-bdb3-746b364256c1\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.160920 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-config-data-default\") pod \"90592895-1eb8-4b3b-bdb3-746b364256c1\" (UID: \"90592895-1eb8-4b3b-bdb3-746b364256c1\") " Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.161274 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "90592895-1eb8-4b3b-bdb3-746b364256c1" (UID: "90592895-1eb8-4b3b-bdb3-746b364256c1"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.161354 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90592895-1eb8-4b3b-bdb3-746b364256c1-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "90592895-1eb8-4b3b-bdb3-746b364256c1" (UID: "90592895-1eb8-4b3b-bdb3-746b364256c1"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.161749 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "90592895-1eb8-4b3b-bdb3-746b364256c1" (UID: "90592895-1eb8-4b3b-bdb3-746b364256c1"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.162049 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "90592895-1eb8-4b3b-bdb3-746b364256c1" (UID: "90592895-1eb8-4b3b-bdb3-746b364256c1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.166349 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90592895-1eb8-4b3b-bdb3-746b364256c1-kube-api-access-gdcfx" (OuterVolumeSpecName: "kube-api-access-gdcfx") pod "90592895-1eb8-4b3b-bdb3-746b364256c1" (UID: "90592895-1eb8-4b3b-bdb3-746b364256c1"). InnerVolumeSpecName "kube-api-access-gdcfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.172910 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "mysql-db") pod "90592895-1eb8-4b3b-bdb3-746b364256c1" (UID: "90592895-1eb8-4b3b-bdb3-746b364256c1"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.198890 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.213358 4943 generic.go:334] "Generic (PLEG): container finished" podID="ba5c9e69-8d95-43ab-aff2-a2a7447c4920" containerID="015400b434d99dc3ac733aab74722f08d32517102fed938f6550057b7bdb1bbe" exitCode=1 Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.213417 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" event={"ID":"ba5c9e69-8d95-43ab-aff2-a2a7447c4920","Type":"ContainerDied","Data":"015400b434d99dc3ac733aab74722f08d32517102fed938f6550057b7bdb1bbe"} Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.214009 4943 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" secret="" err="secret \"galera-openstack-dockercfg-qdzrv\" not found" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.214069 4943 scope.go:117] "RemoveContainer" containerID="015400b434d99dc3ac733aab74722f08d32517102fed938f6550057b7bdb1bbe" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.216594 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.216594 4943 generic.go:334] "Generic (PLEG): container finished" podID="90592895-1eb8-4b3b-bdb3-746b364256c1" containerID="f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286" exitCode=0 Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.216616 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"90592895-1eb8-4b3b-bdb3-746b364256c1","Type":"ContainerDied","Data":"f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286"} Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.216714 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"90592895-1eb8-4b3b-bdb3-746b364256c1","Type":"ContainerDied","Data":"2247fd983dab29b96e207181e6e8d9e928e9d84210cbd750b079d9b503a9bb7d"} Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.216763 4943 scope.go:117] "RemoveContainer" containerID="f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.254149 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.255584 4943 scope.go:117] "RemoveContainer" containerID="b5258258825b305b21c8434353b1f1e0f2f8aa633a7cd6542d026ebcfcf7643a" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.260817 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.262345 4943 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.262414 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/90592895-1eb8-4b3b-bdb3-746b364256c1-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.262487 4943 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.262559 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdcfx\" (UniqueName: \"kubernetes.io/projected/90592895-1eb8-4b3b-bdb3-746b364256c1-kube-api-access-gdcfx\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.262614 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.262667 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/90592895-1eb8-4b3b-bdb3-746b364256c1-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.272911 4943 scope.go:117] "RemoveContainer" containerID="f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286" Dec 04 10:38:06 crc kubenswrapper[4943]: E1204 10:38:06.273405 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286\": container with ID starting with f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286 not found: ID does not exist" containerID="f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.273452 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286"} err="failed to get container status \"f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286\": rpc error: code = NotFound desc = could not find container \"f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286\": container with ID starting with f035ed89589c7a872137cd66eee96a6e4229a448b2e6723a157174738d684286 not found: ID does not exist" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.273475 4943 scope.go:117] "RemoveContainer" containerID="b5258258825b305b21c8434353b1f1e0f2f8aa633a7cd6542d026ebcfcf7643a" Dec 04 10:38:06 crc kubenswrapper[4943]: E1204 10:38:06.273696 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5258258825b305b21c8434353b1f1e0f2f8aa633a7cd6542d026ebcfcf7643a\": container with ID starting with b5258258825b305b21c8434353b1f1e0f2f8aa633a7cd6542d026ebcfcf7643a not found: ID does not exist" containerID="b5258258825b305b21c8434353b1f1e0f2f8aa633a7cd6542d026ebcfcf7643a" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.273722 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5258258825b305b21c8434353b1f1e0f2f8aa633a7cd6542d026ebcfcf7643a"} err="failed to get container status \"b5258258825b305b21c8434353b1f1e0f2f8aa633a7cd6542d026ebcfcf7643a\": rpc error: code = NotFound desc = could not find container \"b5258258825b305b21c8434353b1f1e0f2f8aa633a7cd6542d026ebcfcf7643a\": container with ID starting with b5258258825b305b21c8434353b1f1e0f2f8aa633a7cd6542d026ebcfcf7643a not found: ID does not exist" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.275614 4943 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.364721 4943 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:06 crc kubenswrapper[4943]: E1204 10:38:06.365967 4943 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Dec 04 10:38:06 crc kubenswrapper[4943]: E1204 10:38:06.366832 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-operator-scripts podName:ba5c9e69-8d95-43ab-aff2-a2a7447c4920 nodeName:}" failed. No retries permitted until 2025-12-04 10:38:06.86674947 +0000 UTC m=+1335.455725508 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-operator-scripts") pod "keystonefea1-account-delete-c7hkh" (UID: "ba5c9e69-8d95-43ab-aff2-a2a7447c4920") : configmap "openstack-scripts" not found Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.575665 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90592895-1eb8-4b3b-bdb3-746b364256c1" path="/var/lib/kubelet/pods/90592895-1eb8-4b3b-bdb3-746b364256c1/volumes" Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.576388 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 04 10:38:06 crc kubenswrapper[4943]: I1204 10:38:06.619430 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/rabbitmq-server-0" podUID="057bf926-e2de-427f-9288-345fa9a798a6" containerName="rabbitmq" containerID="cri-o://1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7" gracePeriod=604800 Dec 04 10:38:06 crc kubenswrapper[4943]: E1204 10:38:06.872422 4943 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Dec 04 10:38:06 crc kubenswrapper[4943]: E1204 10:38:06.872572 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-operator-scripts podName:ba5c9e69-8d95-43ab-aff2-a2a7447c4920 nodeName:}" failed. No retries permitted until 2025-12-04 10:38:07.872551216 +0000 UTC m=+1336.461527084 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-operator-scripts") pod "keystonefea1-account-delete-c7hkh" (UID: "ba5c9e69-8d95-43ab-aff2-a2a7447c4920") : configmap "openstack-scripts" not found Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.224813 4943 generic.go:334] "Generic (PLEG): container finished" podID="197663a6-bdeb-49ee-ad21-9b76212842e9" containerID="982cf322919a66576035509dc13c52ef8d2f4017b05fe27ba6629d63ee3e4f1a" exitCode=0 Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.224881 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"197663a6-bdeb-49ee-ad21-9b76212842e9","Type":"ContainerDied","Data":"982cf322919a66576035509dc13c52ef8d2f4017b05fe27ba6629d63ee3e4f1a"} Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.229806 4943 generic.go:334] "Generic (PLEG): container finished" podID="ba5c9e69-8d95-43ab-aff2-a2a7447c4920" containerID="bd267045002f765f964889be265bf7ecc13ae0c723a3ca1e436e40e7c8359f98" exitCode=1 Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.229839 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" event={"ID":"ba5c9e69-8d95-43ab-aff2-a2a7447c4920","Type":"ContainerDied","Data":"bd267045002f765f964889be265bf7ecc13ae0c723a3ca1e436e40e7c8359f98"} Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.229866 4943 scope.go:117] "RemoveContainer" containerID="015400b434d99dc3ac733aab74722f08d32517102fed938f6550057b7bdb1bbe" Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.230458 4943 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" secret="" err="secret \"galera-openstack-dockercfg-qdzrv\" not found" Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.230529 4943 scope.go:117] "RemoveContainer" containerID="bd267045002f765f964889be265bf7ecc13ae0c723a3ca1e436e40e7c8359f98" Dec 04 10:38:07 crc kubenswrapper[4943]: E1204 10:38:07.230880 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystonefea1-account-delete-c7hkh_cinder-kuttl-tests(ba5c9e69-8d95-43ab-aff2-a2a7447c4920)\"" pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" podUID="ba5c9e69-8d95-43ab-aff2-a2a7447c4920" Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.311320 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/openstack-galera-1" podUID="e75cbefc-6761-45f0-86a7-d0ada42d35db" containerName="galera" containerID="cri-o://61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9" gracePeriod=28 Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.563995 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28"] Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.564318 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" podUID="6f3f81a7-15eb-4630-9ccb-f253cb6c7b03" containerName="manager" containerID="cri-o://8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe" gracePeriod=10 Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.636277 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.786646 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/197663a6-bdeb-49ee-ad21-9b76212842e9-config-data\") pod \"197663a6-bdeb-49ee-ad21-9b76212842e9\" (UID: \"197663a6-bdeb-49ee-ad21-9b76212842e9\") " Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.786742 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qdzx\" (UniqueName: \"kubernetes.io/projected/197663a6-bdeb-49ee-ad21-9b76212842e9-kube-api-access-8qdzx\") pod \"197663a6-bdeb-49ee-ad21-9b76212842e9\" (UID: \"197663a6-bdeb-49ee-ad21-9b76212842e9\") " Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.786822 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/197663a6-bdeb-49ee-ad21-9b76212842e9-kolla-config\") pod \"197663a6-bdeb-49ee-ad21-9b76212842e9\" (UID: \"197663a6-bdeb-49ee-ad21-9b76212842e9\") " Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.787742 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/197663a6-bdeb-49ee-ad21-9b76212842e9-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "197663a6-bdeb-49ee-ad21-9b76212842e9" (UID: "197663a6-bdeb-49ee-ad21-9b76212842e9"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.788149 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/197663a6-bdeb-49ee-ad21-9b76212842e9-config-data" (OuterVolumeSpecName: "config-data") pod "197663a6-bdeb-49ee-ad21-9b76212842e9" (UID: "197663a6-bdeb-49ee-ad21-9b76212842e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.806412 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/197663a6-bdeb-49ee-ad21-9b76212842e9-kube-api-access-8qdzx" (OuterVolumeSpecName: "kube-api-access-8qdzx") pod "197663a6-bdeb-49ee-ad21-9b76212842e9" (UID: "197663a6-bdeb-49ee-ad21-9b76212842e9"). InnerVolumeSpecName "kube-api-access-8qdzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.889040 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/197663a6-bdeb-49ee-ad21-9b76212842e9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.889067 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qdzx\" (UniqueName: \"kubernetes.io/projected/197663a6-bdeb-49ee-ad21-9b76212842e9-kube-api-access-8qdzx\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.889076 4943 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/197663a6-bdeb-49ee-ad21-9b76212842e9-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:07 crc kubenswrapper[4943]: E1204 10:38:07.889172 4943 configmap.go:193] Couldn't get configMap cinder-kuttl-tests/openstack-scripts: configmap "openstack-scripts" not found Dec 04 10:38:07 crc kubenswrapper[4943]: E1204 10:38:07.889241 4943 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-operator-scripts podName:ba5c9e69-8d95-43ab-aff2-a2a7447c4920 nodeName:}" failed. No retries permitted until 2025-12-04 10:38:09.889225857 +0000 UTC m=+1338.478201725 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-operator-scripts") pod "keystonefea1-account-delete-c7hkh" (UID: "ba5c9e69-8d95-43ab-aff2-a2a7447c4920") : configmap "openstack-scripts" not found Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.892472 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-index-qdncn"] Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.892686 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/cinder-operator-index-qdncn" podUID="ccd80df8-e6b5-4714-a462-fce5c2b6af85" containerName="registry-server" containerID="cri-o://07400d6f43ac162265f1754d269322529733bda95ac6ecd55d24c4bcdda2e175" gracePeriod=30 Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.926794 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv"] Dec 04 10:38:07 crc kubenswrapper[4943]: I1204 10:38:07.932444 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/ee99d9931aa0f268a41d908d8d03947131e81b1105b4fd2158285b3149jh6pv"] Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.035635 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.047132 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.195228 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zvx8\" (UniqueName: \"kubernetes.io/projected/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-kube-api-access-4zvx8\") pod \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.195297 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-fernet-keys\") pod \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.195337 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-config-data\") pod \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.195353 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4mp7\" (UniqueName: \"kubernetes.io/projected/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-kube-api-access-l4mp7\") pod \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\" (UID: \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.195381 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-credential-keys\") pod \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.195764 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-webhook-cert\") pod \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\" (UID: \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.195812 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-scripts\") pod \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\" (UID: \"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.195846 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-apiservice-cert\") pod \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\" (UID: \"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.199157 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-kube-api-access-l4mp7" (OuterVolumeSpecName: "kube-api-access-l4mp7") pod "6f3f81a7-15eb-4630-9ccb-f253cb6c7b03" (UID: "6f3f81a7-15eb-4630-9ccb-f253cb6c7b03"). InnerVolumeSpecName "kube-api-access-l4mp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.200526 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-scripts" (OuterVolumeSpecName: "scripts") pod "93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a" (UID: "93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.200559 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "6f3f81a7-15eb-4630-9ccb-f253cb6c7b03" (UID: "6f3f81a7-15eb-4630-9ccb-f253cb6c7b03"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.200654 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-kube-api-access-4zvx8" (OuterVolumeSpecName: "kube-api-access-4zvx8") pod "93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a" (UID: "93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a"). InnerVolumeSpecName "kube-api-access-4zvx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.200711 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "6f3f81a7-15eb-4630-9ccb-f253cb6c7b03" (UID: "6f3f81a7-15eb-4630-9ccb-f253cb6c7b03"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.201272 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a" (UID: "93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.204032 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a" (UID: "93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.220860 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-config-data" (OuterVolumeSpecName: "config-data") pod "93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a" (UID: "93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.223524 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.239291 4943 generic.go:334] "Generic (PLEG): container finished" podID="6f3f81a7-15eb-4630-9ccb-f253cb6c7b03" containerID="8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe" exitCode=0 Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.239361 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.239427 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" event={"ID":"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03","Type":"ContainerDied","Data":"8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe"} Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.239463 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28" event={"ID":"6f3f81a7-15eb-4630-9ccb-f253cb6c7b03","Type":"ContainerDied","Data":"c141f7bcd8030f2a01ebc9c515394a6d35466299a120cea8518687a23cd57e49"} Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.239482 4943 scope.go:117] "RemoveContainer" containerID="8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.242387 4943 generic.go:334] "Generic (PLEG): container finished" podID="057bf926-e2de-427f-9288-345fa9a798a6" containerID="1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7" exitCode=0 Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.242457 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"057bf926-e2de-427f-9288-345fa9a798a6","Type":"ContainerDied","Data":"1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7"} Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.242478 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"057bf926-e2de-427f-9288-345fa9a798a6","Type":"ContainerDied","Data":"874b07b83b2f9bd524cdaee4fb66fae48a808c137190f2d0fb3011fdbbc911db"} Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.242557 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.257859 4943 generic.go:334] "Generic (PLEG): container finished" podID="ccd80df8-e6b5-4714-a462-fce5c2b6af85" containerID="07400d6f43ac162265f1754d269322529733bda95ac6ecd55d24c4bcdda2e175" exitCode=0 Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.257969 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-qdncn" event={"ID":"ccd80df8-e6b5-4714-a462-fce5c2b6af85","Type":"ContainerDied","Data":"07400d6f43ac162265f1754d269322529733bda95ac6ecd55d24c4bcdda2e175"} Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.269339 4943 generic.go:334] "Generic (PLEG): container finished" podID="93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a" containerID="5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf" exitCode=0 Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.269447 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" event={"ID":"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a","Type":"ContainerDied","Data":"5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf"} Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.269480 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" event={"ID":"93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a","Type":"ContainerDied","Data":"73e9654e5337260ddea4fab5e4e2306e99511f77d3bedca23ce24a1a5d503efd"} Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.269558 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.280439 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28"] Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.285014 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-5f5d978b7f-vlt28"] Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.288059 4943 scope.go:117] "RemoveContainer" containerID="8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe" Dec 04 10:38:08 crc kubenswrapper[4943]: E1204 10:38:08.289598 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe\": container with ID starting with 8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe not found: ID does not exist" containerID="8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.289641 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe"} err="failed to get container status \"8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe\": rpc error: code = NotFound desc = could not find container \"8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe\": container with ID starting with 8bd2f8560fc8ddfebb989e4a77e171f2b957772523f44236fab0377237cc82fe not found: ID does not exist" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.289672 4943 scope.go:117] "RemoveContainer" containerID="1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.294818 4943 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" secret="" err="secret \"galera-openstack-dockercfg-qdzrv\" not found" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.294871 4943 scope.go:117] "RemoveContainer" containerID="bd267045002f765f964889be265bf7ecc13ae0c723a3ca1e436e40e7c8359f98" Dec 04 10:38:08 crc kubenswrapper[4943]: E1204 10:38:08.295136 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mariadb-account-delete\" with CrashLoopBackOff: \"back-off 10s restarting failed container=mariadb-account-delete pod=keystonefea1-account-delete-c7hkh_cinder-kuttl-tests(ba5c9e69-8d95-43ab-aff2-a2a7447c4920)\"" pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" podUID="ba5c9e69-8d95-43ab-aff2-a2a7447c4920" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.298951 4943 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.298988 4943 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.298997 4943 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.299008 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zvx8\" (UniqueName: \"kubernetes.io/projected/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-kube-api-access-4zvx8\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.299017 4943 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.299025 4943 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.299035 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4mp7\" (UniqueName: \"kubernetes.io/projected/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03-kube-api-access-l4mp7\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.299044 4943 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.311353 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2"] Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.313304 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"197663a6-bdeb-49ee-ad21-9b76212842e9","Type":"ContainerDied","Data":"511d05696b35e326de2476ac9349af3bde12c28d35da0deeabce0f357e038054"} Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.313364 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.319565 4943 scope.go:117] "RemoveContainer" containerID="04af2f59ff3e95f895fbde1e50874174d5ef7263e4b181a70b8a4da6607535f2" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.320105 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-6d8bfb9775-vvbq2"] Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.344308 4943 scope.go:117] "RemoveContainer" containerID="1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7" Dec 04 10:38:08 crc kubenswrapper[4943]: E1204 10:38:08.344888 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7\": container with ID starting with 1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7 not found: ID does not exist" containerID="1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.344941 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7"} err="failed to get container status \"1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7\": rpc error: code = NotFound desc = could not find container \"1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7\": container with ID starting with 1c5106cc032ae3079d6f8135f9e27eaf4e3f34a02471d735afd55416de15e9a7 not found: ID does not exist" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.344979 4943 scope.go:117] "RemoveContainer" containerID="04af2f59ff3e95f895fbde1e50874174d5ef7263e4b181a70b8a4da6607535f2" Dec 04 10:38:08 crc kubenswrapper[4943]: E1204 10:38:08.345723 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04af2f59ff3e95f895fbde1e50874174d5ef7263e4b181a70b8a4da6607535f2\": container with ID starting with 04af2f59ff3e95f895fbde1e50874174d5ef7263e4b181a70b8a4da6607535f2 not found: ID does not exist" containerID="04af2f59ff3e95f895fbde1e50874174d5ef7263e4b181a70b8a4da6607535f2" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.345772 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04af2f59ff3e95f895fbde1e50874174d5ef7263e4b181a70b8a4da6607535f2"} err="failed to get container status \"04af2f59ff3e95f895fbde1e50874174d5ef7263e4b181a70b8a4da6607535f2\": rpc error: code = NotFound desc = could not find container \"04af2f59ff3e95f895fbde1e50874174d5ef7263e4b181a70b8a4da6607535f2\": container with ID starting with 04af2f59ff3e95f895fbde1e50874174d5ef7263e4b181a70b8a4da6607535f2 not found: ID does not exist" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.345810 4943 scope.go:117] "RemoveContainer" containerID="5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.348257 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.354313 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.362826 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-qdncn" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.371389 4943 scope.go:117] "RemoveContainer" containerID="5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf" Dec 04 10:38:08 crc kubenswrapper[4943]: E1204 10:38:08.371889 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf\": container with ID starting with 5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf not found: ID does not exist" containerID="5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.371939 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf"} err="failed to get container status \"5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf\": rpc error: code = NotFound desc = could not find container \"5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf\": container with ID starting with 5ebc63d50a9eed16f4197097f1cc164032f3c89a370103429c725dcc9c468faf not found: ID does not exist" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.371973 4943 scope.go:117] "RemoveContainer" containerID="982cf322919a66576035509dc13c52ef8d2f4017b05fe27ba6629d63ee3e4f1a" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.400373 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-confd\") pod \"057bf926-e2de-427f-9288-345fa9a798a6\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.400708 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/057bf926-e2de-427f-9288-345fa9a798a6-erlang-cookie-secret\") pod \"057bf926-e2de-427f-9288-345fa9a798a6\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.400785 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/057bf926-e2de-427f-9288-345fa9a798a6-pod-info\") pod \"057bf926-e2de-427f-9288-345fa9a798a6\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.400857 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpx9s\" (UniqueName: \"kubernetes.io/projected/057bf926-e2de-427f-9288-345fa9a798a6-kube-api-access-tpx9s\") pod \"057bf926-e2de-427f-9288-345fa9a798a6\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.400903 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/057bf926-e2de-427f-9288-345fa9a798a6-plugins-conf\") pod \"057bf926-e2de-427f-9288-345fa9a798a6\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.401116 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7afe53b7-993a-480f-963a-183d18cbdf12\") pod \"057bf926-e2de-427f-9288-345fa9a798a6\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.401172 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-erlang-cookie\") pod \"057bf926-e2de-427f-9288-345fa9a798a6\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.401236 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-plugins\") pod \"057bf926-e2de-427f-9288-345fa9a798a6\" (UID: \"057bf926-e2de-427f-9288-345fa9a798a6\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.402777 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "057bf926-e2de-427f-9288-345fa9a798a6" (UID: "057bf926-e2de-427f-9288-345fa9a798a6"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.403829 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/057bf926-e2de-427f-9288-345fa9a798a6-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "057bf926-e2de-427f-9288-345fa9a798a6" (UID: "057bf926-e2de-427f-9288-345fa9a798a6"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.403940 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057bf926-e2de-427f-9288-345fa9a798a6-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "057bf926-e2de-427f-9288-345fa9a798a6" (UID: "057bf926-e2de-427f-9288-345fa9a798a6"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.405150 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "057bf926-e2de-427f-9288-345fa9a798a6" (UID: "057bf926-e2de-427f-9288-345fa9a798a6"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.406675 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/057bf926-e2de-427f-9288-345fa9a798a6-pod-info" (OuterVolumeSpecName: "pod-info") pod "057bf926-e2de-427f-9288-345fa9a798a6" (UID: "057bf926-e2de-427f-9288-345fa9a798a6"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.408436 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/057bf926-e2de-427f-9288-345fa9a798a6-kube-api-access-tpx9s" (OuterVolumeSpecName: "kube-api-access-tpx9s") pod "057bf926-e2de-427f-9288-345fa9a798a6" (UID: "057bf926-e2de-427f-9288-345fa9a798a6"). InnerVolumeSpecName "kube-api-access-tpx9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.414146 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7afe53b7-993a-480f-963a-183d18cbdf12" (OuterVolumeSpecName: "persistence") pod "057bf926-e2de-427f-9288-345fa9a798a6" (UID: "057bf926-e2de-427f-9288-345fa9a798a6"). InnerVolumeSpecName "pvc-7afe53b7-993a-480f-963a-183d18cbdf12". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.461476 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "057bf926-e2de-427f-9288-345fa9a798a6" (UID: "057bf926-e2de-427f-9288-345fa9a798a6"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.502889 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpz7t\" (UniqueName: \"kubernetes.io/projected/ccd80df8-e6b5-4714-a462-fce5c2b6af85-kube-api-access-gpz7t\") pod \"ccd80df8-e6b5-4714-a462-fce5c2b6af85\" (UID: \"ccd80df8-e6b5-4714-a462-fce5c2b6af85\") " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.503359 4943 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.503387 4943 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.503408 4943 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/057bf926-e2de-427f-9288-345fa9a798a6-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.503421 4943 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/057bf926-e2de-427f-9288-345fa9a798a6-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.503434 4943 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/057bf926-e2de-427f-9288-345fa9a798a6-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.503446 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpx9s\" (UniqueName: \"kubernetes.io/projected/057bf926-e2de-427f-9288-345fa9a798a6-kube-api-access-tpx9s\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.503457 4943 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/057bf926-e2de-427f-9288-345fa9a798a6-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.503516 4943 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-7afe53b7-993a-480f-963a-183d18cbdf12\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7afe53b7-993a-480f-963a-183d18cbdf12\") on node \"crc\" " Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.506396 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccd80df8-e6b5-4714-a462-fce5c2b6af85-kube-api-access-gpz7t" (OuterVolumeSpecName: "kube-api-access-gpz7t") pod "ccd80df8-e6b5-4714-a462-fce5c2b6af85" (UID: "ccd80df8-e6b5-4714-a462-fce5c2b6af85"). InnerVolumeSpecName "kube-api-access-gpz7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.516093 4943 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.516232 4943 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-7afe53b7-993a-480f-963a-183d18cbdf12" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7afe53b7-993a-480f-963a-183d18cbdf12") on node "crc" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.579570 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="197663a6-bdeb-49ee-ad21-9b76212842e9" path="/var/lib/kubelet/pods/197663a6-bdeb-49ee-ad21-9b76212842e9/volumes" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.580268 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3349476a-f5bc-40ce-a663-561662c8a6ac" path="/var/lib/kubelet/pods/3349476a-f5bc-40ce-a663-561662c8a6ac/volumes" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.581023 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f3f81a7-15eb-4630-9ccb-f253cb6c7b03" path="/var/lib/kubelet/pods/6f3f81a7-15eb-4630-9ccb-f253cb6c7b03/volumes" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.583082 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a" path="/var/lib/kubelet/pods/93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a/volumes" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.589519 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.591458 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.604482 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpz7t\" (UniqueName: \"kubernetes.io/projected/ccd80df8-e6b5-4714-a462-fce5c2b6af85-kube-api-access-gpz7t\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:08 crc kubenswrapper[4943]: I1204 10:38:08.604522 4943 reconciler_common.go:293] "Volume detached for volume \"pvc-7afe53b7-993a-480f-963a-183d18cbdf12\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7afe53b7-993a-480f-963a-183d18cbdf12\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.192740 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.314446 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-operator-scripts\") pod \"e75cbefc-6761-45f0-86a7-d0ada42d35db\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.314498 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-config-data-default\") pod \"e75cbefc-6761-45f0-86a7-d0ada42d35db\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.314584 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"e75cbefc-6761-45f0-86a7-d0ada42d35db\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.314639 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-kolla-config\") pod \"e75cbefc-6761-45f0-86a7-d0ada42d35db\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.314667 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdwmr\" (UniqueName: \"kubernetes.io/projected/e75cbefc-6761-45f0-86a7-d0ada42d35db-kube-api-access-bdwmr\") pod \"e75cbefc-6761-45f0-86a7-d0ada42d35db\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.314741 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e75cbefc-6761-45f0-86a7-d0ada42d35db-config-data-generated\") pod \"e75cbefc-6761-45f0-86a7-d0ada42d35db\" (UID: \"e75cbefc-6761-45f0-86a7-d0ada42d35db\") " Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.315384 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e75cbefc-6761-45f0-86a7-d0ada42d35db-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "e75cbefc-6761-45f0-86a7-d0ada42d35db" (UID: "e75cbefc-6761-45f0-86a7-d0ada42d35db"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.315582 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e75cbefc-6761-45f0-86a7-d0ada42d35db" (UID: "e75cbefc-6761-45f0-86a7-d0ada42d35db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.315643 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "e75cbefc-6761-45f0-86a7-d0ada42d35db" (UID: "e75cbefc-6761-45f0-86a7-d0ada42d35db"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.316070 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "e75cbefc-6761-45f0-86a7-d0ada42d35db" (UID: "e75cbefc-6761-45f0-86a7-d0ada42d35db"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.327506 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e75cbefc-6761-45f0-86a7-d0ada42d35db-kube-api-access-bdwmr" (OuterVolumeSpecName: "kube-api-access-bdwmr") pod "e75cbefc-6761-45f0-86a7-d0ada42d35db" (UID: "e75cbefc-6761-45f0-86a7-d0ada42d35db"). InnerVolumeSpecName "kube-api-access-bdwmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.329855 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/openstack-galera-0" podUID="55950271-9dac-4a16-8cca-5288ddb40dba" containerName="galera" containerID="cri-o://9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc" gracePeriod=26 Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.330801 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-qdncn" event={"ID":"ccd80df8-e6b5-4714-a462-fce5c2b6af85","Type":"ContainerDied","Data":"559a16573294063b207a10d73ec01316e7a7d688aacbec9e5d06ace47e8b4aa6"} Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.330849 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-qdncn" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.331024 4943 scope.go:117] "RemoveContainer" containerID="07400d6f43ac162265f1754d269322529733bda95ac6ecd55d24c4bcdda2e175" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.332667 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "mysql-db") pod "e75cbefc-6761-45f0-86a7-d0ada42d35db" (UID: "e75cbefc-6761-45f0-86a7-d0ada42d35db"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.344874 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-dndb4"] Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.346489 4943 generic.go:334] "Generic (PLEG): container finished" podID="e75cbefc-6761-45f0-86a7-d0ada42d35db" containerID="61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9" exitCode=0 Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.346749 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"e75cbefc-6761-45f0-86a7-d0ada42d35db","Type":"ContainerDied","Data":"61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9"} Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.346782 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"e75cbefc-6761-45f0-86a7-d0ada42d35db","Type":"ContainerDied","Data":"5569a85e24f4743cecf769628abfb8142d82ce9c24a58caadcae27b1f52d1a6a"} Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.346902 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.351649 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-dndb4"] Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.367293 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j"] Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.379807 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystonefea1-account-delete-c7hkh"] Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.392294 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-fea1-account-create-update-gpj8j"] Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.392821 4943 scope.go:117] "RemoveContainer" containerID="61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.400841 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/cinder-operator-index-qdncn"] Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.404738 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/cinder-operator-index-qdncn"] Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.409651 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.417267 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.417720 4943 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.417748 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdwmr\" (UniqueName: \"kubernetes.io/projected/e75cbefc-6761-45f0-86a7-d0ada42d35db-kube-api-access-bdwmr\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.417763 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e75cbefc-6761-45f0-86a7-d0ada42d35db-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.417774 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.417786 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e75cbefc-6761-45f0-86a7-d0ada42d35db-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.417811 4943 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.418644 4943 scope.go:117] "RemoveContainer" containerID="0e58835abd4cbec14ccbc6596bb0acee42ba89f53d70bca8690701acfd3aee55" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.431733 4943 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.462512 4943 scope.go:117] "RemoveContainer" containerID="61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9" Dec 04 10:38:09 crc kubenswrapper[4943]: E1204 10:38:09.471652 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9\": container with ID starting with 61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9 not found: ID does not exist" containerID="61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.471703 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9"} err="failed to get container status \"61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9\": rpc error: code = NotFound desc = could not find container \"61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9\": container with ID starting with 61887b1107239647e07bdcf282eccd91815be498cd9f66f858a46aaf1d54c1b9 not found: ID does not exist" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.471737 4943 scope.go:117] "RemoveContainer" containerID="0e58835abd4cbec14ccbc6596bb0acee42ba89f53d70bca8690701acfd3aee55" Dec 04 10:38:09 crc kubenswrapper[4943]: E1204 10:38:09.477843 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e58835abd4cbec14ccbc6596bb0acee42ba89f53d70bca8690701acfd3aee55\": container with ID starting with 0e58835abd4cbec14ccbc6596bb0acee42ba89f53d70bca8690701acfd3aee55 not found: ID does not exist" containerID="0e58835abd4cbec14ccbc6596bb0acee42ba89f53d70bca8690701acfd3aee55" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.477883 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e58835abd4cbec14ccbc6596bb0acee42ba89f53d70bca8690701acfd3aee55"} err="failed to get container status \"0e58835abd4cbec14ccbc6596bb0acee42ba89f53d70bca8690701acfd3aee55\": rpc error: code = NotFound desc = could not find container \"0e58835abd4cbec14ccbc6596bb0acee42ba89f53d70bca8690701acfd3aee55\": container with ID starting with 0e58835abd4cbec14ccbc6596bb0acee42ba89f53d70bca8690701acfd3aee55 not found: ID does not exist" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.518762 4943 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.592967 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.620406 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-operator-scripts\") pod \"ba5c9e69-8d95-43ab-aff2-a2a7447c4920\" (UID: \"ba5c9e69-8d95-43ab-aff2-a2a7447c4920\") " Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.621252 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ba5c9e69-8d95-43ab-aff2-a2a7447c4920" (UID: "ba5c9e69-8d95-43ab-aff2-a2a7447c4920"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.721939 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jnf7\" (UniqueName: \"kubernetes.io/projected/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-kube-api-access-6jnf7\") pod \"ba5c9e69-8d95-43ab-aff2-a2a7447c4920\" (UID: \"ba5c9e69-8d95-43ab-aff2-a2a7447c4920\") " Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.722457 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.726319 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-kube-api-access-6jnf7" (OuterVolumeSpecName: "kube-api-access-6jnf7") pod "ba5c9e69-8d95-43ab-aff2-a2a7447c4920" (UID: "ba5c9e69-8d95-43ab-aff2-a2a7447c4920"). InnerVolumeSpecName "kube-api-access-6jnf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:09 crc kubenswrapper[4943]: I1204 10:38:09.825017 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jnf7\" (UniqueName: \"kubernetes.io/projected/ba5c9e69-8d95-43ab-aff2-a2a7447c4920-kube-api-access-6jnf7\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.011245 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7"] Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.011488 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" podUID="f6006a85-81c2-4909-912e-16c49c862f7a" containerName="manager" containerID="cri-o://7bd46b939209cfe6a74549112d929c167c3310fc4fc3f47953a37260e2b66125" gracePeriod=10 Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.172249 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.230328 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-operator-scripts\") pod \"55950271-9dac-4a16-8cca-5288ddb40dba\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.230367 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"55950271-9dac-4a16-8cca-5288ddb40dba\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.230422 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/55950271-9dac-4a16-8cca-5288ddb40dba-config-data-generated\") pod \"55950271-9dac-4a16-8cca-5288ddb40dba\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.230457 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-kolla-config\") pod \"55950271-9dac-4a16-8cca-5288ddb40dba\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.230485 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrd6b\" (UniqueName: \"kubernetes.io/projected/55950271-9dac-4a16-8cca-5288ddb40dba-kube-api-access-vrd6b\") pod \"55950271-9dac-4a16-8cca-5288ddb40dba\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.230517 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-config-data-default\") pod \"55950271-9dac-4a16-8cca-5288ddb40dba\" (UID: \"55950271-9dac-4a16-8cca-5288ddb40dba\") " Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.231124 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "55950271-9dac-4a16-8cca-5288ddb40dba" (UID: "55950271-9dac-4a16-8cca-5288ddb40dba"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.231855 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "55950271-9dac-4a16-8cca-5288ddb40dba" (UID: "55950271-9dac-4a16-8cca-5288ddb40dba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.232956 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55950271-9dac-4a16-8cca-5288ddb40dba-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "55950271-9dac-4a16-8cca-5288ddb40dba" (UID: "55950271-9dac-4a16-8cca-5288ddb40dba"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.233880 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "55950271-9dac-4a16-8cca-5288ddb40dba" (UID: "55950271-9dac-4a16-8cca-5288ddb40dba"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.238352 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55950271-9dac-4a16-8cca-5288ddb40dba-kube-api-access-vrd6b" (OuterVolumeSpecName: "kube-api-access-vrd6b") pod "55950271-9dac-4a16-8cca-5288ddb40dba" (UID: "55950271-9dac-4a16-8cca-5288ddb40dba"). InnerVolumeSpecName "kube-api-access-vrd6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.250107 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "55950271-9dac-4a16-8cca-5288ddb40dba" (UID: "55950271-9dac-4a16-8cca-5288ddb40dba"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.269463 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-p744p"] Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.269681 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-index-p744p" podUID="f5acada6-99e9-4109-b6fe-0b636e85a756" containerName="registry-server" containerID="cri-o://3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6" gracePeriod=30 Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.295248 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp"] Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.307404 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/49c083020ae5dfe237b73a6c6b807501660a323f061d7879268c43a121vkggp"] Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.331912 4943 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.331971 4943 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.331982 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/55950271-9dac-4a16-8cca-5288ddb40dba-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.331993 4943 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.332001 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrd6b\" (UniqueName: \"kubernetes.io/projected/55950271-9dac-4a16-8cca-5288ddb40dba-kube-api-access-vrd6b\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.332011 4943 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/55950271-9dac-4a16-8cca-5288ddb40dba-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.342299 4943 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.368356 4943 generic.go:334] "Generic (PLEG): container finished" podID="f6006a85-81c2-4909-912e-16c49c862f7a" containerID="7bd46b939209cfe6a74549112d929c167c3310fc4fc3f47953a37260e2b66125" exitCode=0 Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.368461 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" event={"ID":"f6006a85-81c2-4909-912e-16c49c862f7a","Type":"ContainerDied","Data":"7bd46b939209cfe6a74549112d929c167c3310fc4fc3f47953a37260e2b66125"} Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.372685 4943 generic.go:334] "Generic (PLEG): container finished" podID="55950271-9dac-4a16-8cca-5288ddb40dba" containerID="9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc" exitCode=0 Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.372752 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"55950271-9dac-4a16-8cca-5288ddb40dba","Type":"ContainerDied","Data":"9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc"} Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.372775 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"55950271-9dac-4a16-8cca-5288ddb40dba","Type":"ContainerDied","Data":"52c85e45d636b61dce12cef4dcf398c3860441e0964b2dcfeb2d06a0bdd1ce70"} Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.372797 4943 scope.go:117] "RemoveContainer" containerID="9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.372906 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.388659 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" event={"ID":"ba5c9e69-8d95-43ab-aff2-a2a7447c4920","Type":"ContainerDied","Data":"743549dc69b994c64fff097be4a625dd0ddf654116af8581d4a64411ed140e7e"} Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.388776 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystonefea1-account-delete-c7hkh" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.427962 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.435088 4943 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.446394 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.446901 4943 scope.go:117] "RemoveContainer" containerID="3e2a8a68745c611b3c975bedaa991c7e4881c6bcb3a24bf94d91e2ba5b91a7e4" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.454332 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.471627 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystonefea1-account-delete-c7hkh"] Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.476145 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystonefea1-account-delete-c7hkh"] Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.476806 4943 scope.go:117] "RemoveContainer" containerID="9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc" Dec 04 10:38:10 crc kubenswrapper[4943]: E1204 10:38:10.477305 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc\": container with ID starting with 9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc not found: ID does not exist" containerID="9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.477369 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc"} err="failed to get container status \"9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc\": rpc error: code = NotFound desc = could not find container \"9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc\": container with ID starting with 9f9d2df25c60b6c079ca38de75f6c35512e855586b665f75c21e115fc581b3fc not found: ID does not exist" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.477410 4943 scope.go:117] "RemoveContainer" containerID="3e2a8a68745c611b3c975bedaa991c7e4881c6bcb3a24bf94d91e2ba5b91a7e4" Dec 04 10:38:10 crc kubenswrapper[4943]: E1204 10:38:10.478288 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e2a8a68745c611b3c975bedaa991c7e4881c6bcb3a24bf94d91e2ba5b91a7e4\": container with ID starting with 3e2a8a68745c611b3c975bedaa991c7e4881c6bcb3a24bf94d91e2ba5b91a7e4 not found: ID does not exist" containerID="3e2a8a68745c611b3c975bedaa991c7e4881c6bcb3a24bf94d91e2ba5b91a7e4" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.478329 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e2a8a68745c611b3c975bedaa991c7e4881c6bcb3a24bf94d91e2ba5b91a7e4"} err="failed to get container status \"3e2a8a68745c611b3c975bedaa991c7e4881c6bcb3a24bf94d91e2ba5b91a7e4\": rpc error: code = NotFound desc = could not find container \"3e2a8a68745c611b3c975bedaa991c7e4881c6bcb3a24bf94d91e2ba5b91a7e4\": container with ID starting with 3e2a8a68745c611b3c975bedaa991c7e4881c6bcb3a24bf94d91e2ba5b91a7e4 not found: ID does not exist" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.478363 4943 scope.go:117] "RemoveContainer" containerID="bd267045002f765f964889be265bf7ecc13ae0c723a3ca1e436e40e7c8359f98" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.535674 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6006a85-81c2-4909-912e-16c49c862f7a-webhook-cert\") pod \"f6006a85-81c2-4909-912e-16c49c862f7a\" (UID: \"f6006a85-81c2-4909-912e-16c49c862f7a\") " Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.535731 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6drc\" (UniqueName: \"kubernetes.io/projected/f6006a85-81c2-4909-912e-16c49c862f7a-kube-api-access-z6drc\") pod \"f6006a85-81c2-4909-912e-16c49c862f7a\" (UID: \"f6006a85-81c2-4909-912e-16c49c862f7a\") " Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.535807 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6006a85-81c2-4909-912e-16c49c862f7a-apiservice-cert\") pod \"f6006a85-81c2-4909-912e-16c49c862f7a\" (UID: \"f6006a85-81c2-4909-912e-16c49c862f7a\") " Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.539421 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6006a85-81c2-4909-912e-16c49c862f7a-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "f6006a85-81c2-4909-912e-16c49c862f7a" (UID: "f6006a85-81c2-4909-912e-16c49c862f7a"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.548429 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6006a85-81c2-4909-912e-16c49c862f7a-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "f6006a85-81c2-4909-912e-16c49c862f7a" (UID: "f6006a85-81c2-4909-912e-16c49c862f7a"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.548462 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6006a85-81c2-4909-912e-16c49c862f7a-kube-api-access-z6drc" (OuterVolumeSpecName: "kube-api-access-z6drc") pod "f6006a85-81c2-4909-912e-16c49c862f7a" (UID: "f6006a85-81c2-4909-912e-16c49c862f7a"). InnerVolumeSpecName "kube-api-access-z6drc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.579529 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="057bf926-e2de-427f-9288-345fa9a798a6" path="/var/lib/kubelet/pods/057bf926-e2de-427f-9288-345fa9a798a6/volumes" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.580093 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2391487b-52c7-4e1f-a03b-b20371f30445" path="/var/lib/kubelet/pods/2391487b-52c7-4e1f-a03b-b20371f30445/volumes" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.581275 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55950271-9dac-4a16-8cca-5288ddb40dba" path="/var/lib/kubelet/pods/55950271-9dac-4a16-8cca-5288ddb40dba/volumes" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.581846 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="954e673d-85ec-4ebd-9c7a-105986c8ccab" path="/var/lib/kubelet/pods/954e673d-85ec-4ebd-9c7a-105986c8ccab/volumes" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.582382 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6266a3a-7624-4cd4-b9de-fb806517a35c" path="/var/lib/kubelet/pods/a6266a3a-7624-4cd4-b9de-fb806517a35c/volumes" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.582861 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba5c9e69-8d95-43ab-aff2-a2a7447c4920" path="/var/lib/kubelet/pods/ba5c9e69-8d95-43ab-aff2-a2a7447c4920/volumes" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.584124 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccd80df8-e6b5-4714-a462-fce5c2b6af85" path="/var/lib/kubelet/pods/ccd80df8-e6b5-4714-a462-fce5c2b6af85/volumes" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.584794 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e75cbefc-6761-45f0-86a7-d0ada42d35db" path="/var/lib/kubelet/pods/e75cbefc-6761-45f0-86a7-d0ada42d35db/volumes" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.637594 4943 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f6006a85-81c2-4909-912e-16c49c862f7a-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.637914 4943 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f6006a85-81c2-4909-912e-16c49c862f7a-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.637929 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6drc\" (UniqueName: \"kubernetes.io/projected/f6006a85-81c2-4909-912e-16c49c862f7a-kube-api-access-z6drc\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.649362 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-p744p" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.841058 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6vz9\" (UniqueName: \"kubernetes.io/projected/f5acada6-99e9-4109-b6fe-0b636e85a756-kube-api-access-p6vz9\") pod \"f5acada6-99e9-4109-b6fe-0b636e85a756\" (UID: \"f5acada6-99e9-4109-b6fe-0b636e85a756\") " Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.845322 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5acada6-99e9-4109-b6fe-0b636e85a756-kube-api-access-p6vz9" (OuterVolumeSpecName: "kube-api-access-p6vz9") pod "f5acada6-99e9-4109-b6fe-0b636e85a756" (UID: "f5acada6-99e9-4109-b6fe-0b636e85a756"). InnerVolumeSpecName "kube-api-access-p6vz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:10 crc kubenswrapper[4943]: I1204 10:38:10.943315 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6vz9\" (UniqueName: \"kubernetes.io/projected/f5acada6-99e9-4109-b6fe-0b636e85a756-kube-api-access-p6vz9\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.399858 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" event={"ID":"f6006a85-81c2-4909-912e-16c49c862f7a","Type":"ContainerDied","Data":"296c2f2b7582287bb4102cf997c45adb71085a3e41dcebf80012c6b2005aa4fc"} Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.399888 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7" Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.399920 4943 scope.go:117] "RemoveContainer" containerID="7bd46b939209cfe6a74549112d929c167c3310fc4fc3f47953a37260e2b66125" Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.406541 4943 generic.go:334] "Generic (PLEG): container finished" podID="f5acada6-99e9-4109-b6fe-0b636e85a756" containerID="3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6" exitCode=0 Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.406622 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-p744p" event={"ID":"f5acada6-99e9-4109-b6fe-0b636e85a756","Type":"ContainerDied","Data":"3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6"} Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.406653 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-p744p" event={"ID":"f5acada6-99e9-4109-b6fe-0b636e85a756","Type":"ContainerDied","Data":"97227c3e16bdc36f994f5f805ddf9fd10fb69469b9539838c257329919d2317a"} Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.406730 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-p744p" Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.418099 4943 scope.go:117] "RemoveContainer" containerID="3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6" Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.421420 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7"] Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.428517 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-794886f574-lzdr7"] Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.437610 4943 scope.go:117] "RemoveContainer" containerID="3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6" Dec 04 10:38:11 crc kubenswrapper[4943]: E1204 10:38:11.438014 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6\": container with ID starting with 3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6 not found: ID does not exist" containerID="3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6" Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.438047 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6"} err="failed to get container status \"3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6\": rpc error: code = NotFound desc = could not find container \"3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6\": container with ID starting with 3cf0f9eea38997df589cc9975e3aca5b5390a4b77094889ffe85386a226ff2a6 not found: ID does not exist" Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.452063 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/keystone-operator-index-p744p"] Dec 04 10:38:11 crc kubenswrapper[4943]: I1204 10:38:11.457404 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/keystone-operator-index-p744p"] Dec 04 10:38:12 crc kubenswrapper[4943]: I1204 10:38:12.573843 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5acada6-99e9-4109-b6fe-0b636e85a756" path="/var/lib/kubelet/pods/f5acada6-99e9-4109-b6fe-0b636e85a756/volumes" Dec 04 10:38:12 crc kubenswrapper[4943]: I1204 10:38:12.574427 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6006a85-81c2-4909-912e-16c49c862f7a" path="/var/lib/kubelet/pods/f6006a85-81c2-4909-912e-16c49c862f7a/volumes" Dec 04 10:38:13 crc kubenswrapper[4943]: I1204 10:38:13.959044 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97"] Dec 04 10:38:13 crc kubenswrapper[4943]: I1204 10:38:13.959465 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" podUID="9cbafb93-24ac-4d39-bd39-0869654fb225" containerName="operator" containerID="cri-o://f42b9950e81784930b83aced47ed9ea7c6642570c940c3d6eb86bb6ce8ab6115" gracePeriod=10 Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.318509 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-s99cd"] Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.319628 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" podUID="34001aba-ee9d-4c56-9895-009f4fea5926" containerName="registry-server" containerID="cri-o://2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59" gracePeriod=30 Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.354092 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk"] Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.374058 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590845rk"] Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.434354 4943 generic.go:334] "Generic (PLEG): container finished" podID="9cbafb93-24ac-4d39-bd39-0869654fb225" containerID="f42b9950e81784930b83aced47ed9ea7c6642570c940c3d6eb86bb6ce8ab6115" exitCode=0 Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.434411 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" event={"ID":"9cbafb93-24ac-4d39-bd39-0869654fb225","Type":"ContainerDied","Data":"f42b9950e81784930b83aced47ed9ea7c6642570c940c3d6eb86bb6ce8ab6115"} Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.434493 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" event={"ID":"9cbafb93-24ac-4d39-bd39-0869654fb225","Type":"ContainerDied","Data":"3f6066fb3bd1d4fce48e876f041587fd7a498ca0ebc986b96a0dc5dd74acadb2"} Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.434541 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f6066fb3bd1d4fce48e876f041587fd7a498ca0ebc986b96a0dc5dd74acadb2" Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.491335 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.496853 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c98l\" (UniqueName: \"kubernetes.io/projected/9cbafb93-24ac-4d39-bd39-0869654fb225-kube-api-access-7c98l\") pod \"9cbafb93-24ac-4d39-bd39-0869654fb225\" (UID: \"9cbafb93-24ac-4d39-bd39-0869654fb225\") " Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.503163 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cbafb93-24ac-4d39-bd39-0869654fb225-kube-api-access-7c98l" (OuterVolumeSpecName: "kube-api-access-7c98l") pod "9cbafb93-24ac-4d39-bd39-0869654fb225" (UID: "9cbafb93-24ac-4d39-bd39-0869654fb225"). InnerVolumeSpecName "kube-api-access-7c98l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.576537 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e9104f1-6500-4ca1-88a4-b7b912c51caa" path="/var/lib/kubelet/pods/7e9104f1-6500-4ca1-88a4-b7b912c51caa/volumes" Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.598173 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c98l\" (UniqueName: \"kubernetes.io/projected/9cbafb93-24ac-4d39-bd39-0869654fb225-kube-api-access-7c98l\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.637579 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.698530 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvph2\" (UniqueName: \"kubernetes.io/projected/34001aba-ee9d-4c56-9895-009f4fea5926-kube-api-access-kvph2\") pod \"34001aba-ee9d-4c56-9895-009f4fea5926\" (UID: \"34001aba-ee9d-4c56-9895-009f4fea5926\") " Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.702600 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34001aba-ee9d-4c56-9895-009f4fea5926-kube-api-access-kvph2" (OuterVolumeSpecName: "kube-api-access-kvph2") pod "34001aba-ee9d-4c56-9895-009f4fea5926" (UID: "34001aba-ee9d-4c56-9895-009f4fea5926"). InnerVolumeSpecName "kube-api-access-kvph2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:14 crc kubenswrapper[4943]: I1204 10:38:14.799670 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvph2\" (UniqueName: \"kubernetes.io/projected/34001aba-ee9d-4c56-9895-009f4fea5926-kube-api-access-kvph2\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:15 crc kubenswrapper[4943]: I1204 10:38:15.442310 4943 generic.go:334] "Generic (PLEG): container finished" podID="34001aba-ee9d-4c56-9895-009f4fea5926" containerID="2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59" exitCode=0 Dec 04 10:38:15 crc kubenswrapper[4943]: I1204 10:38:15.442388 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" event={"ID":"34001aba-ee9d-4c56-9895-009f4fea5926","Type":"ContainerDied","Data":"2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59"} Dec 04 10:38:15 crc kubenswrapper[4943]: I1204 10:38:15.442683 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97" Dec 04 10:38:15 crc kubenswrapper[4943]: I1204 10:38:15.442491 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" Dec 04 10:38:15 crc kubenswrapper[4943]: I1204 10:38:15.442757 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-s99cd" event={"ID":"34001aba-ee9d-4c56-9895-009f4fea5926","Type":"ContainerDied","Data":"ff3ec7c7a11cc2e2d43ddddf298724e44b197d6ceb8d3016c2b54bce5178e49d"} Dec 04 10:38:15 crc kubenswrapper[4943]: I1204 10:38:15.443186 4943 scope.go:117] "RemoveContainer" containerID="2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59" Dec 04 10:38:15 crc kubenswrapper[4943]: I1204 10:38:15.459423 4943 scope.go:117] "RemoveContainer" containerID="2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59" Dec 04 10:38:15 crc kubenswrapper[4943]: E1204 10:38:15.459938 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59\": container with ID starting with 2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59 not found: ID does not exist" containerID="2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59" Dec 04 10:38:15 crc kubenswrapper[4943]: I1204 10:38:15.459995 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59"} err="failed to get container status \"2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59\": rpc error: code = NotFound desc = could not find container \"2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59\": container with ID starting with 2c8a605c5882ce1587fe7ebc3b10ce0c0f68bcaf411338023a9eba963692fe59 not found: ID does not exist" Dec 04 10:38:15 crc kubenswrapper[4943]: I1204 10:38:15.468909 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-s99cd"] Dec 04 10:38:15 crc kubenswrapper[4943]: I1204 10:38:15.476385 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-s99cd"] Dec 04 10:38:15 crc kubenswrapper[4943]: I1204 10:38:15.481581 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97"] Dec 04 10:38:15 crc kubenswrapper[4943]: I1204 10:38:15.487093 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-gsn97"] Dec 04 10:38:16 crc kubenswrapper[4943]: I1204 10:38:16.573320 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34001aba-ee9d-4c56-9895-009f4fea5926" path="/var/lib/kubelet/pods/34001aba-ee9d-4c56-9895-009f4fea5926/volumes" Dec 04 10:38:16 crc kubenswrapper[4943]: I1204 10:38:16.573909 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cbafb93-24ac-4d39-bd39-0869654fb225" path="/var/lib/kubelet/pods/9cbafb93-24ac-4d39-bd39-0869654fb225/volumes" Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.233698 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8"] Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.234037 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" podUID="5e7c05b7-ea2c-4771-ac23-f2217871d2b3" containerName="manager" containerID="cri-o://6442c1ee2235ed93b4bda925ce0d21931ebcb953ac8c2184b305eedc4dfe77a2" gracePeriod=10 Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.234103 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" podUID="5e7c05b7-ea2c-4771-ac23-f2217871d2b3" containerName="kube-rbac-proxy" containerID="cri-o://724e34ee83c66974dc2a6c83b4871b4fa49659cace9f6f61b91581129c990efb" gracePeriod=10 Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.458142 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-szxbw"] Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.458770 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/infra-operator-index-szxbw" podUID="56e01120-0537-47f3-92d0-c9b6ebc9651a" containerName="registry-server" containerID="cri-o://5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f" gracePeriod=30 Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.513101 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd"] Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.513469 4943 generic.go:334] "Generic (PLEG): container finished" podID="5e7c05b7-ea2c-4771-ac23-f2217871d2b3" containerID="724e34ee83c66974dc2a6c83b4871b4fa49659cace9f6f61b91581129c990efb" exitCode=0 Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.513494 4943 generic.go:334] "Generic (PLEG): container finished" podID="5e7c05b7-ea2c-4771-ac23-f2217871d2b3" containerID="6442c1ee2235ed93b4bda925ce0d21931ebcb953ac8c2184b305eedc4dfe77a2" exitCode=0 Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.513509 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" event={"ID":"5e7c05b7-ea2c-4771-ac23-f2217871d2b3","Type":"ContainerDied","Data":"724e34ee83c66974dc2a6c83b4871b4fa49659cace9f6f61b91581129c990efb"} Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.513532 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" event={"ID":"5e7c05b7-ea2c-4771-ac23-f2217871d2b3","Type":"ContainerDied","Data":"6442c1ee2235ed93b4bda925ce0d21931ebcb953ac8c2184b305eedc4dfe77a2"} Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.518510 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dbqznd"] Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.572225 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21b39ac1-11f1-4bf2-bdd8-6e93062ecafe" path="/var/lib/kubelet/pods/21b39ac1-11f1-4bf2-bdd8-6e93062ecafe/volumes" Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.642084 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.642182 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.682960 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.855603 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgd9s\" (UniqueName: \"kubernetes.io/projected/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-kube-api-access-dgd9s\") pod \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\" (UID: \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\") " Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.855662 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-webhook-cert\") pod \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\" (UID: \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\") " Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.855691 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-apiservice-cert\") pod \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\" (UID: \"5e7c05b7-ea2c-4771-ac23-f2217871d2b3\") " Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.858953 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-szxbw" Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.860730 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-kube-api-access-dgd9s" (OuterVolumeSpecName: "kube-api-access-dgd9s") pod "5e7c05b7-ea2c-4771-ac23-f2217871d2b3" (UID: "5e7c05b7-ea2c-4771-ac23-f2217871d2b3"). InnerVolumeSpecName "kube-api-access-dgd9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.860699 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "5e7c05b7-ea2c-4771-ac23-f2217871d2b3" (UID: "5e7c05b7-ea2c-4771-ac23-f2217871d2b3"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.860992 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "5e7c05b7-ea2c-4771-ac23-f2217871d2b3" (UID: "5e7c05b7-ea2c-4771-ac23-f2217871d2b3"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.957237 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgd9s\" (UniqueName: \"kubernetes.io/projected/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-kube-api-access-dgd9s\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.957445 4943 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:18 crc kubenswrapper[4943]: I1204 10:38:18.957521 4943 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e7c05b7-ea2c-4771-ac23-f2217871d2b3-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.058072 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5xqp\" (UniqueName: \"kubernetes.io/projected/56e01120-0537-47f3-92d0-c9b6ebc9651a-kube-api-access-w5xqp\") pod \"56e01120-0537-47f3-92d0-c9b6ebc9651a\" (UID: \"56e01120-0537-47f3-92d0-c9b6ebc9651a\") " Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.060849 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56e01120-0537-47f3-92d0-c9b6ebc9651a-kube-api-access-w5xqp" (OuterVolumeSpecName: "kube-api-access-w5xqp") pod "56e01120-0537-47f3-92d0-c9b6ebc9651a" (UID: "56e01120-0537-47f3-92d0-c9b6ebc9651a"). InnerVolumeSpecName "kube-api-access-w5xqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.159096 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5xqp\" (UniqueName: \"kubernetes.io/projected/56e01120-0537-47f3-92d0-c9b6ebc9651a-kube-api-access-w5xqp\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.520501 4943 generic.go:334] "Generic (PLEG): container finished" podID="56e01120-0537-47f3-92d0-c9b6ebc9651a" containerID="5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f" exitCode=0 Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.520589 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-szxbw" Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.520620 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-szxbw" event={"ID":"56e01120-0537-47f3-92d0-c9b6ebc9651a","Type":"ContainerDied","Data":"5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f"} Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.521017 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-szxbw" event={"ID":"56e01120-0537-47f3-92d0-c9b6ebc9651a","Type":"ContainerDied","Data":"771be2b6914a587f236deb123a86642711698a7330a333bee1a982e41d1e2e67"} Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.521062 4943 scope.go:117] "RemoveContainer" containerID="5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f" Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.523773 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" event={"ID":"5e7c05b7-ea2c-4771-ac23-f2217871d2b3","Type":"ContainerDied","Data":"236d3612fb0c2f98b72196abd9ea9aa30cf4650f23fd66b90a25fa6acf698fe9"} Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.523871 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8" Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.536278 4943 scope.go:117] "RemoveContainer" containerID="5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f" Dec 04 10:38:19 crc kubenswrapper[4943]: E1204 10:38:19.536992 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f\": container with ID starting with 5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f not found: ID does not exist" containerID="5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f" Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.537039 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f"} err="failed to get container status \"5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f\": rpc error: code = NotFound desc = could not find container \"5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f\": container with ID starting with 5b6ec8c60242ce0a629749866d96c146e1b6f05afc88687699a3b17512011a3f not found: ID does not exist" Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.537059 4943 scope.go:117] "RemoveContainer" containerID="724e34ee83c66974dc2a6c83b4871b4fa49659cace9f6f61b91581129c990efb" Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.550915 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-index-szxbw"] Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.557494 4943 scope.go:117] "RemoveContainer" containerID="6442c1ee2235ed93b4bda925ce0d21931ebcb953ac8c2184b305eedc4dfe77a2" Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.565443 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-index-szxbw"] Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.570670 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8"] Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.574765 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/infra-operator-controller-manager-db7f8fc9c-qsqx8"] Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.816399 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7"] Dec 04 10:38:19 crc kubenswrapper[4943]: I1204 10:38:19.816887 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" podUID="10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0" containerName="manager" containerID="cri-o://afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f" gracePeriod=10 Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.157982 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-q9s4x"] Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.158238 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-q9s4x" podUID="c076efe8-6293-4cd9-a27d-d6196e44b0e2" containerName="registry-server" containerID="cri-o://6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f" gracePeriod=30 Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.180835 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7"] Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.188564 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/55a3886cc1ed42812df4eab61c7a6033dc924d195539e8545c8f175f61pfng7"] Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.296991 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.377519 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-webhook-cert\") pod \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\" (UID: \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\") " Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.377641 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5t6r\" (UniqueName: \"kubernetes.io/projected/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-kube-api-access-l5t6r\") pod \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\" (UID: \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\") " Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.377672 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-apiservice-cert\") pod \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\" (UID: \"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0\") " Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.384453 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0" (UID: "10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.388449 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0" (UID: "10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.390401 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-kube-api-access-l5t6r" (OuterVolumeSpecName: "kube-api-access-l5t6r") pod "10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0" (UID: "10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0"). InnerVolumeSpecName "kube-api-access-l5t6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.478820 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5t6r\" (UniqueName: \"kubernetes.io/projected/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-kube-api-access-l5t6r\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.478859 4943 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.478870 4943 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.526568 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-q9s4x" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.532250 4943 generic.go:334] "Generic (PLEG): container finished" podID="c076efe8-6293-4cd9-a27d-d6196e44b0e2" containerID="6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f" exitCode=0 Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.532316 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-q9s4x" event={"ID":"c076efe8-6293-4cd9-a27d-d6196e44b0e2","Type":"ContainerDied","Data":"6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f"} Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.532347 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-q9s4x" event={"ID":"c076efe8-6293-4cd9-a27d-d6196e44b0e2","Type":"ContainerDied","Data":"c837524f810ec82eed9b7b785475b2cf2f134fb1fcbc5721d20ad55a30479cc1"} Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.532368 4943 scope.go:117] "RemoveContainer" containerID="6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.532449 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-q9s4x" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.535158 4943 generic.go:334] "Generic (PLEG): container finished" podID="10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0" containerID="afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f" exitCode=0 Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.535253 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" event={"ID":"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0","Type":"ContainerDied","Data":"afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f"} Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.535294 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" event={"ID":"10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0","Type":"ContainerDied","Data":"21322eb6f4d98f29386068d4fdc46fb089cc0ff5dc166c109affb7a6c3e45828"} Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.535328 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.553589 4943 scope.go:117] "RemoveContainer" containerID="6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f" Dec 04 10:38:20 crc kubenswrapper[4943]: E1204 10:38:20.554019 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f\": container with ID starting with 6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f not found: ID does not exist" containerID="6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.554040 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f"} err="failed to get container status \"6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f\": rpc error: code = NotFound desc = could not find container \"6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f\": container with ID starting with 6be4e239dbf56359efa276255853230e818c10c9c613abcc6f5615855322933f not found: ID does not exist" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.554060 4943 scope.go:117] "RemoveContainer" containerID="afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.574257 4943 scope.go:117] "RemoveContainer" containerID="afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f" Dec 04 10:38:20 crc kubenswrapper[4943]: E1204 10:38:20.574845 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f\": container with ID starting with afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f not found: ID does not exist" containerID="afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.574916 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f"} err="failed to get container status \"afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f\": rpc error: code = NotFound desc = could not find container \"afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f\": container with ID starting with afd6b4c395153797fa86c9adeea8610dbc5392c7d7fb8c8c57dd674400c1b47f not found: ID does not exist" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.579464 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbdh8\" (UniqueName: \"kubernetes.io/projected/c076efe8-6293-4cd9-a27d-d6196e44b0e2-kube-api-access-gbdh8\") pod \"c076efe8-6293-4cd9-a27d-d6196e44b0e2\" (UID: \"c076efe8-6293-4cd9-a27d-d6196e44b0e2\") " Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.581003 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56e01120-0537-47f3-92d0-c9b6ebc9651a" path="/var/lib/kubelet/pods/56e01120-0537-47f3-92d0-c9b6ebc9651a/volumes" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.581543 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e7c05b7-ea2c-4771-ac23-f2217871d2b3" path="/var/lib/kubelet/pods/5e7c05b7-ea2c-4771-ac23-f2217871d2b3/volumes" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.602911 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ec088ed-72ed-4c49-bba9-53047c3bf094" path="/var/lib/kubelet/pods/5ec088ed-72ed-4c49-bba9-53047c3bf094/volumes" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.607176 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c076efe8-6293-4cd9-a27d-d6196e44b0e2-kube-api-access-gbdh8" (OuterVolumeSpecName: "kube-api-access-gbdh8") pod "c076efe8-6293-4cd9-a27d-d6196e44b0e2" (UID: "c076efe8-6293-4cd9-a27d-d6196e44b0e2"). InnerVolumeSpecName "kube-api-access-gbdh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.626411 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7"] Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.626460 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-84c9b4cc96-2frr7"] Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.680457 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbdh8\" (UniqueName: \"kubernetes.io/projected/c076efe8-6293-4cd9-a27d-d6196e44b0e2-kube-api-access-gbdh8\") on node \"crc\" DevicePath \"\"" Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.857566 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-q9s4x"] Dec 04 10:38:20 crc kubenswrapper[4943]: I1204 10:38:20.862903 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-q9s4x"] Dec 04 10:38:22 crc kubenswrapper[4943]: I1204 10:38:22.572527 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0" path="/var/lib/kubelet/pods/10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0/volumes" Dec 04 10:38:22 crc kubenswrapper[4943]: I1204 10:38:22.572970 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c076efe8-6293-4cd9-a27d-d6196e44b0e2" path="/var/lib/kubelet/pods/c076efe8-6293-4cd9-a27d-d6196e44b0e2/volumes" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.558795 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-hjzgj/must-gather-sscvm"] Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559714 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a" containerName="keystone-api" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559748 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a" containerName="keystone-api" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559761 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c076efe8-6293-4cd9-a27d-d6196e44b0e2" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559770 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="c076efe8-6293-4cd9-a27d-d6196e44b0e2" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559779 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5acada6-99e9-4109-b6fe-0b636e85a756" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559786 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5acada6-99e9-4109-b6fe-0b636e85a756" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559802 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="197663a6-bdeb-49ee-ad21-9b76212842e9" containerName="memcached" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559810 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="197663a6-bdeb-49ee-ad21-9b76212842e9" containerName="memcached" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559820 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7c05b7-ea2c-4771-ac23-f2217871d2b3" containerName="manager" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559828 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7c05b7-ea2c-4771-ac23-f2217871d2b3" containerName="manager" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559837 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7c05b7-ea2c-4771-ac23-f2217871d2b3" containerName="kube-rbac-proxy" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559844 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7c05b7-ea2c-4771-ac23-f2217871d2b3" containerName="kube-rbac-proxy" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559860 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55950271-9dac-4a16-8cca-5288ddb40dba" containerName="galera" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559867 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="55950271-9dac-4a16-8cca-5288ddb40dba" containerName="galera" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559883 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0" containerName="manager" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559890 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0" containerName="manager" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559898 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56e01120-0537-47f3-92d0-c9b6ebc9651a" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559905 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="56e01120-0537-47f3-92d0-c9b6ebc9651a" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559919 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057bf926-e2de-427f-9288-345fa9a798a6" containerName="rabbitmq" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559926 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="057bf926-e2de-427f-9288-345fa9a798a6" containerName="rabbitmq" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559938 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34001aba-ee9d-4c56-9895-009f4fea5926" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559944 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="34001aba-ee9d-4c56-9895-009f4fea5926" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559955 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057bf926-e2de-427f-9288-345fa9a798a6" containerName="setup-container" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559963 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="057bf926-e2de-427f-9288-345fa9a798a6" containerName="setup-container" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559975 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90592895-1eb8-4b3b-bdb3-746b364256c1" containerName="mysql-bootstrap" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559982 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="90592895-1eb8-4b3b-bdb3-746b364256c1" containerName="mysql-bootstrap" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.559992 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6006a85-81c2-4909-912e-16c49c862f7a" containerName="manager" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.559999 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6006a85-81c2-4909-912e-16c49c862f7a" containerName="manager" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.560013 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e75cbefc-6761-45f0-86a7-d0ada42d35db" containerName="galera" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560023 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e75cbefc-6761-45f0-86a7-d0ada42d35db" containerName="galera" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.560033 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerName="cinder-volume" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560041 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerName="cinder-volume" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.560053 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e75cbefc-6761-45f0-86a7-d0ada42d35db" containerName="mysql-bootstrap" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560060 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="e75cbefc-6761-45f0-86a7-d0ada42d35db" containerName="mysql-bootstrap" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.560069 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55950271-9dac-4a16-8cca-5288ddb40dba" containerName="mysql-bootstrap" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560076 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="55950271-9dac-4a16-8cca-5288ddb40dba" containerName="mysql-bootstrap" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.560088 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cbafb93-24ac-4d39-bd39-0869654fb225" containerName="operator" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560095 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cbafb93-24ac-4d39-bd39-0869654fb225" containerName="operator" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.560103 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90592895-1eb8-4b3b-bdb3-746b364256c1" containerName="galera" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560110 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="90592895-1eb8-4b3b-bdb3-746b364256c1" containerName="galera" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.560120 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5c9e69-8d95-43ab-aff2-a2a7447c4920" containerName="mariadb-account-delete" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560127 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5c9e69-8d95-43ab-aff2-a2a7447c4920" containerName="mariadb-account-delete" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.560138 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f3f81a7-15eb-4630-9ccb-f253cb6c7b03" containerName="manager" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560146 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f3f81a7-15eb-4630-9ccb-f253cb6c7b03" containerName="manager" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.560158 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5c9e69-8d95-43ab-aff2-a2a7447c4920" containerName="mariadb-account-delete" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560167 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5c9e69-8d95-43ab-aff2-a2a7447c4920" containerName="mariadb-account-delete" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.560180 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd80df8-e6b5-4714-a462-fce5c2b6af85" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560187 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd80df8-e6b5-4714-a462-fce5c2b6af85" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: E1204 10:38:34.560202 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerName="probe" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560211 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerName="probe" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560396 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerName="cinder-volume" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560409 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="10935c2c-5d3a-43ae-9f98-cc5b0a1a1da0" containerName="manager" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560417 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5acada6-99e9-4109-b6fe-0b636e85a756" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560426 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="55950271-9dac-4a16-8cca-5288ddb40dba" containerName="galera" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560435 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="90592895-1eb8-4b3b-bdb3-746b364256c1" containerName="galera" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560441 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="197663a6-bdeb-49ee-ad21-9b76212842e9" containerName="memcached" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560450 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cbafb93-24ac-4d39-bd39-0869654fb225" containerName="operator" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560457 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e7c05b7-ea2c-4771-ac23-f2217871d2b3" containerName="kube-rbac-proxy" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560463 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="93e2e3e2-5464-4a4a-a0aa-81c4d2d2ad3a" containerName="keystone-api" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560470 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e7c05b7-ea2c-4771-ac23-f2217871d2b3" containerName="manager" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560478 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="c076efe8-6293-4cd9-a27d-d6196e44b0e2" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560486 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="34001aba-ee9d-4c56-9895-009f4fea5926" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560494 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba5c9e69-8d95-43ab-aff2-a2a7447c4920" containerName="mariadb-account-delete" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560502 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba5c9e69-8d95-43ab-aff2-a2a7447c4920" containerName="mariadb-account-delete" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560510 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6006a85-81c2-4909-912e-16c49c862f7a" containerName="manager" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560516 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="057bf926-e2de-427f-9288-345fa9a798a6" containerName="rabbitmq" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560524 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b1200a-9080-4647-bb43-844e4cd27f6d" containerName="probe" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560531 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="e75cbefc-6761-45f0-86a7-d0ada42d35db" containerName="galera" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560538 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd80df8-e6b5-4714-a462-fce5c2b6af85" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560545 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f3f81a7-15eb-4630-9ccb-f253cb6c7b03" containerName="manager" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.560552 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="56e01120-0537-47f3-92d0-c9b6ebc9651a" containerName="registry-server" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.561275 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjzgj/must-gather-sscvm" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.564285 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hjzgj"/"openshift-service-ca.crt" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.564314 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-hjzgj"/"default-dockercfg-ffplf" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.564505 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-hjzgj"/"kube-root-ca.crt" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.572950 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hjzgj/must-gather-sscvm"] Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.738393 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q99v\" (UniqueName: \"kubernetes.io/projected/53981522-c74a-43ce-be48-271982fe9485-kube-api-access-6q99v\") pod \"must-gather-sscvm\" (UID: \"53981522-c74a-43ce-be48-271982fe9485\") " pod="openshift-must-gather-hjzgj/must-gather-sscvm" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.738467 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/53981522-c74a-43ce-be48-271982fe9485-must-gather-output\") pod \"must-gather-sscvm\" (UID: \"53981522-c74a-43ce-be48-271982fe9485\") " pod="openshift-must-gather-hjzgj/must-gather-sscvm" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.839870 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q99v\" (UniqueName: \"kubernetes.io/projected/53981522-c74a-43ce-be48-271982fe9485-kube-api-access-6q99v\") pod \"must-gather-sscvm\" (UID: \"53981522-c74a-43ce-be48-271982fe9485\") " pod="openshift-must-gather-hjzgj/must-gather-sscvm" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.839929 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/53981522-c74a-43ce-be48-271982fe9485-must-gather-output\") pod \"must-gather-sscvm\" (UID: \"53981522-c74a-43ce-be48-271982fe9485\") " pod="openshift-must-gather-hjzgj/must-gather-sscvm" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.840425 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/53981522-c74a-43ce-be48-271982fe9485-must-gather-output\") pod \"must-gather-sscvm\" (UID: \"53981522-c74a-43ce-be48-271982fe9485\") " pod="openshift-must-gather-hjzgj/must-gather-sscvm" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.857092 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q99v\" (UniqueName: \"kubernetes.io/projected/53981522-c74a-43ce-be48-271982fe9485-kube-api-access-6q99v\") pod \"must-gather-sscvm\" (UID: \"53981522-c74a-43ce-be48-271982fe9485\") " pod="openshift-must-gather-hjzgj/must-gather-sscvm" Dec 04 10:38:34 crc kubenswrapper[4943]: I1204 10:38:34.926124 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjzgj/must-gather-sscvm" Dec 04 10:38:35 crc kubenswrapper[4943]: I1204 10:38:35.310382 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-hjzgj/must-gather-sscvm"] Dec 04 10:38:35 crc kubenswrapper[4943]: W1204 10:38:35.317336 4943 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53981522_c74a_43ce_be48_271982fe9485.slice/crio-00b5f79d03c54791f373c4c4f353542a500800eca110cd0f3efbddad110aec1a WatchSource:0}: Error finding container 00b5f79d03c54791f373c4c4f353542a500800eca110cd0f3efbddad110aec1a: Status 404 returned error can't find the container with id 00b5f79d03c54791f373c4c4f353542a500800eca110cd0f3efbddad110aec1a Dec 04 10:38:35 crc kubenswrapper[4943]: I1204 10:38:35.717149 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjzgj/must-gather-sscvm" event={"ID":"53981522-c74a-43ce-be48-271982fe9485","Type":"ContainerStarted","Data":"00b5f79d03c54791f373c4c4f353542a500800eca110cd0f3efbddad110aec1a"} Dec 04 10:38:39 crc kubenswrapper[4943]: I1204 10:38:39.747254 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjzgj/must-gather-sscvm" event={"ID":"53981522-c74a-43ce-be48-271982fe9485","Type":"ContainerStarted","Data":"48740a0a3f8cf0d1df14659c143ad2d919bd2050b5dac5b9960ddea795f75a1d"} Dec 04 10:38:39 crc kubenswrapper[4943]: I1204 10:38:39.747761 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjzgj/must-gather-sscvm" event={"ID":"53981522-c74a-43ce-be48-271982fe9485","Type":"ContainerStarted","Data":"f4ecac3eb0af694a09914f36ddacf9c5f70401b5e8b41b6a0822cf0946de9c53"} Dec 04 10:38:39 crc kubenswrapper[4943]: I1204 10:38:39.763304 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-hjzgj/must-gather-sscvm" podStartSLOduration=2.193613143 podStartE2EDuration="5.763262829s" podCreationTimestamp="2025-12-04 10:38:34 +0000 UTC" firstStartedPulling="2025-12-04 10:38:35.319354202 +0000 UTC m=+1363.908330070" lastFinishedPulling="2025-12-04 10:38:38.889003888 +0000 UTC m=+1367.477979756" observedRunningTime="2025-12-04 10:38:39.759173806 +0000 UTC m=+1368.348149674" watchObservedRunningTime="2025-12-04 10:38:39.763262829 +0000 UTC m=+1368.352238697" Dec 04 10:38:48 crc kubenswrapper[4943]: I1204 10:38:48.629165 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:38:48 crc kubenswrapper[4943]: I1204 10:38:48.629771 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:38:48 crc kubenswrapper[4943]: I1204 10:38:48.629823 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:38:48 crc kubenswrapper[4943]: I1204 10:38:48.630503 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e58a73db13bfe564778718e69895f4217bb13940aaf3f0a781ae48e378dfd34"} pod="openshift-machine-config-operator/machine-config-daemon-kswzd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 10:38:48 crc kubenswrapper[4943]: I1204 10:38:48.630574 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" containerID="cri-o://7e58a73db13bfe564778718e69895f4217bb13940aaf3f0a781ae48e378dfd34" gracePeriod=600 Dec 04 10:38:49 crc kubenswrapper[4943]: I1204 10:38:49.814759 4943 generic.go:334] "Generic (PLEG): container finished" podID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerID="7e58a73db13bfe564778718e69895f4217bb13940aaf3f0a781ae48e378dfd34" exitCode=0 Dec 04 10:38:49 crc kubenswrapper[4943]: I1204 10:38:49.814853 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerDied","Data":"7e58a73db13bfe564778718e69895f4217bb13940aaf3f0a781ae48e378dfd34"} Dec 04 10:38:49 crc kubenswrapper[4943]: I1204 10:38:49.815126 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerStarted","Data":"a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520"} Dec 04 10:38:49 crc kubenswrapper[4943]: I1204 10:38:49.815152 4943 scope.go:117] "RemoveContainer" containerID="39ef936964dcabd0fd14176e3d0e5709046494c4af42c583afefd4c04502ae7a" Dec 04 10:38:53 crc kubenswrapper[4943]: I1204 10:38:53.156241 4943 scope.go:117] "RemoveContainer" containerID="f8d98b25f4434d0a16b3ee37103940be8fff2ad1f838e3c4f5bc113dc67b3b14" Dec 04 10:38:53 crc kubenswrapper[4943]: I1204 10:38:53.175763 4943 scope.go:117] "RemoveContainer" containerID="18c2e1103834644ee7013475e9f5e7580ba6c9948857a4ef3c560c850bf7c70f" Dec 04 10:38:53 crc kubenswrapper[4943]: I1204 10:38:53.196174 4943 scope.go:117] "RemoveContainer" containerID="f0e6f2b6bd6291af317000838c52b98396866b3d0ad06a75314077430f652864" Dec 04 10:38:53 crc kubenswrapper[4943]: I1204 10:38:53.220175 4943 scope.go:117] "RemoveContainer" containerID="3e062fbd9e2ccd9b2c36ca08fd0945e54a464bd590527267dbccd2d508c6f1ea" Dec 04 10:38:53 crc kubenswrapper[4943]: I1204 10:38:53.235195 4943 scope.go:117] "RemoveContainer" containerID="58725c2bb3322200f2389af365db07f9329eda286b3c4c33901f7fc2b03df294" Dec 04 10:38:53 crc kubenswrapper[4943]: I1204 10:38:53.257569 4943 scope.go:117] "RemoveContainer" containerID="6843515e1dead6038699866c979534ba43c1d0a8a3332005df7c21090dbdd13c" Dec 04 10:39:20 crc kubenswrapper[4943]: I1204 10:39:20.039795 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-vhxkd_5ef5c158-a074-47df-a4bb-2e1cab760745/control-plane-machine-set-operator/0.log" Dec 04 10:39:20 crc kubenswrapper[4943]: I1204 10:39:20.172418 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8sdsb_1f2720fe-737e-4562-9fc1-d0e52e0d99a2/kube-rbac-proxy/0.log" Dec 04 10:39:20 crc kubenswrapper[4943]: I1204 10:39:20.245869 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8sdsb_1f2720fe-737e-4562-9fc1-d0e52e0d99a2/machine-api-operator/0.log" Dec 04 10:39:34 crc kubenswrapper[4943]: I1204 10:39:34.955363 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-hm4b7_71342ec6-c2cb-4277-a567-e13aabab1db3/kube-rbac-proxy/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.010831 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-hm4b7_71342ec6-c2cb-4277-a567-e13aabab1db3/controller/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.147473 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-frr-files/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.297394 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-metrics/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.332102 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-frr-files/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.343076 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-reloader/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.362686 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-reloader/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.497912 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-frr-files/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.536589 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-metrics/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.577835 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-reloader/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.591590 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-metrics/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.704521 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-frr-files/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.722097 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-reloader/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.783775 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-metrics/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.790998 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/controller/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.932923 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/frr-metrics/0.log" Dec 04 10:39:35 crc kubenswrapper[4943]: I1204 10:39:35.978083 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/kube-rbac-proxy/0.log" Dec 04 10:39:36 crc kubenswrapper[4943]: I1204 10:39:36.016654 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/kube-rbac-proxy-frr/0.log" Dec 04 10:39:36 crc kubenswrapper[4943]: I1204 10:39:36.141883 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/reloader/0.log" Dec 04 10:39:36 crc kubenswrapper[4943]: I1204 10:39:36.176938 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/frr/0.log" Dec 04 10:39:36 crc kubenswrapper[4943]: I1204 10:39:36.194588 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-t89n4_6104b8e7-773f-4873-912f-5269972dfac9/frr-k8s-webhook-server/0.log" Dec 04 10:39:36 crc kubenswrapper[4943]: I1204 10:39:36.350234 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5697dbbbd-r8zls_0ca89ccd-b04f-480d-9a4e-a96b534b3ca6/manager/0.log" Dec 04 10:39:36 crc kubenswrapper[4943]: I1204 10:39:36.383831 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-57ffff8ff7-k42kt_4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716/webhook-server/0.log" Dec 04 10:39:36 crc kubenswrapper[4943]: I1204 10:39:36.556171 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sd2np_b473a6fb-38de-4766-b414-56f3c0496ff6/kube-rbac-proxy/0.log" Dec 04 10:39:36 crc kubenswrapper[4943]: I1204 10:39:36.703481 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sd2np_b473a6fb-38de-4766-b414-56f3c0496ff6/speaker/0.log" Dec 04 10:39:53 crc kubenswrapper[4943]: I1204 10:39:53.490984 4943 scope.go:117] "RemoveContainer" containerID="3ea468f3c2ce28512325954b214388854cdae92e4b9a990e314f27613d83be1f" Dec 04 10:39:53 crc kubenswrapper[4943]: I1204 10:39:53.525045 4943 scope.go:117] "RemoveContainer" containerID="258b8d02268fdb04b1901f1cdb2afb5e9e91515232a922f604a70c11976b76bc" Dec 04 10:39:53 crc kubenswrapper[4943]: I1204 10:39:53.542277 4943 scope.go:117] "RemoveContainer" containerID="f42b9950e81784930b83aced47ed9ea7c6642570c940c3d6eb86bb6ce8ab6115" Dec 04 10:39:53 crc kubenswrapper[4943]: I1204 10:39:53.563306 4943 scope.go:117] "RemoveContainer" containerID="d6149d6d408f5420971882aac7f59dd83734dcb21f29c76b5b83e3ab43a23f6e" Dec 04 10:39:59 crc kubenswrapper[4943]: I1204 10:39:59.736722 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/util/0.log" Dec 04 10:39:59 crc kubenswrapper[4943]: I1204 10:39:59.930328 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/pull/0.log" Dec 04 10:39:59 crc kubenswrapper[4943]: I1204 10:39:59.942503 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/util/0.log" Dec 04 10:39:59 crc kubenswrapper[4943]: I1204 10:39:59.954321 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/pull/0.log" Dec 04 10:40:00 crc kubenswrapper[4943]: I1204 10:40:00.154388 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/pull/0.log" Dec 04 10:40:00 crc kubenswrapper[4943]: I1204 10:40:00.163093 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/extract/0.log" Dec 04 10:40:00 crc kubenswrapper[4943]: I1204 10:40:00.202323 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/util/0.log" Dec 04 10:40:00 crc kubenswrapper[4943]: I1204 10:40:00.458403 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/extract-utilities/0.log" Dec 04 10:40:00 crc kubenswrapper[4943]: I1204 10:40:00.627494 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/extract-utilities/0.log" Dec 04 10:40:00 crc kubenswrapper[4943]: I1204 10:40:00.633616 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/extract-content/0.log" Dec 04 10:40:00 crc kubenswrapper[4943]: I1204 10:40:00.665900 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/extract-content/0.log" Dec 04 10:40:00 crc kubenswrapper[4943]: I1204 10:40:00.788958 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/extract-utilities/0.log" Dec 04 10:40:00 crc kubenswrapper[4943]: I1204 10:40:00.812887 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/extract-content/0.log" Dec 04 10:40:00 crc kubenswrapper[4943]: I1204 10:40:00.966938 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/extract-utilities/0.log" Dec 04 10:40:01 crc kubenswrapper[4943]: I1204 10:40:01.057364 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/registry-server/0.log" Dec 04 10:40:01 crc kubenswrapper[4943]: I1204 10:40:01.186228 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/extract-content/0.log" Dec 04 10:40:01 crc kubenswrapper[4943]: I1204 10:40:01.191062 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/extract-content/0.log" Dec 04 10:40:01 crc kubenswrapper[4943]: I1204 10:40:01.206774 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/extract-utilities/0.log" Dec 04 10:40:01 crc kubenswrapper[4943]: I1204 10:40:01.362096 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/extract-utilities/0.log" Dec 04 10:40:01 crc kubenswrapper[4943]: I1204 10:40:01.424843 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/extract-content/0.log" Dec 04 10:40:01 crc kubenswrapper[4943]: I1204 10:40:01.612000 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-bwkhz_04c5f3f6-69da-4c89-801f-5e3d519c80ae/marketplace-operator/0.log" Dec 04 10:40:01 crc kubenswrapper[4943]: I1204 10:40:01.681153 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/extract-utilities/0.log" Dec 04 10:40:01 crc kubenswrapper[4943]: I1204 10:40:01.734658 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/registry-server/0.log" Dec 04 10:40:01 crc kubenswrapper[4943]: I1204 10:40:01.867740 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/extract-content/0.log" Dec 04 10:40:01 crc kubenswrapper[4943]: I1204 10:40:01.875677 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/extract-content/0.log" Dec 04 10:40:01 crc kubenswrapper[4943]: I1204 10:40:01.881459 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/extract-utilities/0.log" Dec 04 10:40:02 crc kubenswrapper[4943]: I1204 10:40:02.021262 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/extract-utilities/0.log" Dec 04 10:40:02 crc kubenswrapper[4943]: I1204 10:40:02.024620 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/extract-content/0.log" Dec 04 10:40:02 crc kubenswrapper[4943]: I1204 10:40:02.103313 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/registry-server/0.log" Dec 04 10:40:02 crc kubenswrapper[4943]: I1204 10:40:02.220191 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/extract-utilities/0.log" Dec 04 10:40:02 crc kubenswrapper[4943]: I1204 10:40:02.345792 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/extract-content/0.log" Dec 04 10:40:02 crc kubenswrapper[4943]: I1204 10:40:02.349272 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/extract-utilities/0.log" Dec 04 10:40:02 crc kubenswrapper[4943]: I1204 10:40:02.406238 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/extract-content/0.log" Dec 04 10:40:02 crc kubenswrapper[4943]: I1204 10:40:02.537847 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/extract-content/0.log" Dec 04 10:40:02 crc kubenswrapper[4943]: I1204 10:40:02.545505 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/extract-utilities/0.log" Dec 04 10:40:02 crc kubenswrapper[4943]: I1204 10:40:02.858001 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/registry-server/0.log" Dec 04 10:40:39 crc kubenswrapper[4943]: I1204 10:40:39.976173 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p4glq"] Dec 04 10:40:39 crc kubenswrapper[4943]: I1204 10:40:39.993954 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:40 crc kubenswrapper[4943]: I1204 10:40:39.996995 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p4glq"] Dec 04 10:40:40 crc kubenswrapper[4943]: I1204 10:40:40.001557 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-catalog-content\") pod \"redhat-operators-p4glq\" (UID: \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\") " pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:40 crc kubenswrapper[4943]: I1204 10:40:40.001637 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-utilities\") pod \"redhat-operators-p4glq\" (UID: \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\") " pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:40 crc kubenswrapper[4943]: I1204 10:40:40.001759 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm5kr\" (UniqueName: \"kubernetes.io/projected/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-kube-api-access-bm5kr\") pod \"redhat-operators-p4glq\" (UID: \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\") " pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:40 crc kubenswrapper[4943]: I1204 10:40:40.102837 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm5kr\" (UniqueName: \"kubernetes.io/projected/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-kube-api-access-bm5kr\") pod \"redhat-operators-p4glq\" (UID: \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\") " pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:40 crc kubenswrapper[4943]: I1204 10:40:40.102959 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-catalog-content\") pod \"redhat-operators-p4glq\" (UID: \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\") " pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:40 crc kubenswrapper[4943]: I1204 10:40:40.102988 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-utilities\") pod \"redhat-operators-p4glq\" (UID: \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\") " pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:40 crc kubenswrapper[4943]: I1204 10:40:40.103671 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-utilities\") pod \"redhat-operators-p4glq\" (UID: \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\") " pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:40 crc kubenswrapper[4943]: I1204 10:40:40.103776 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-catalog-content\") pod \"redhat-operators-p4glq\" (UID: \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\") " pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:40 crc kubenswrapper[4943]: I1204 10:40:40.126456 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm5kr\" (UniqueName: \"kubernetes.io/projected/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-kube-api-access-bm5kr\") pod \"redhat-operators-p4glq\" (UID: \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\") " pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:40 crc kubenswrapper[4943]: I1204 10:40:40.328836 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:40 crc kubenswrapper[4943]: I1204 10:40:40.560187 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p4glq"] Dec 04 10:40:41 crc kubenswrapper[4943]: I1204 10:40:41.554350 4943 generic.go:334] "Generic (PLEG): container finished" podID="1d9232b8-b9f5-43f9-b31e-eff6c189d91b" containerID="15d8c3671937450205f3f3c069810ffc189ed3a98e4949675411104bf6452f1d" exitCode=0 Dec 04 10:40:41 crc kubenswrapper[4943]: I1204 10:40:41.554679 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4glq" event={"ID":"1d9232b8-b9f5-43f9-b31e-eff6c189d91b","Type":"ContainerDied","Data":"15d8c3671937450205f3f3c069810ffc189ed3a98e4949675411104bf6452f1d"} Dec 04 10:40:41 crc kubenswrapper[4943]: I1204 10:40:41.554732 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4glq" event={"ID":"1d9232b8-b9f5-43f9-b31e-eff6c189d91b","Type":"ContainerStarted","Data":"fb648bcd221c6ab2921f073cbf3ddad46d18e1f72291f27bca9dc5948a8cbb2e"} Dec 04 10:40:41 crc kubenswrapper[4943]: I1204 10:40:41.556713 4943 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 10:40:42 crc kubenswrapper[4943]: I1204 10:40:42.562281 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4glq" event={"ID":"1d9232b8-b9f5-43f9-b31e-eff6c189d91b","Type":"ContainerStarted","Data":"32291466750bcbd135477c34808983884cc87a42aaa782b65fd71f3ea4ce3c07"} Dec 04 10:40:43 crc kubenswrapper[4943]: I1204 10:40:43.571474 4943 generic.go:334] "Generic (PLEG): container finished" podID="1d9232b8-b9f5-43f9-b31e-eff6c189d91b" containerID="32291466750bcbd135477c34808983884cc87a42aaa782b65fd71f3ea4ce3c07" exitCode=0 Dec 04 10:40:43 crc kubenswrapper[4943]: I1204 10:40:43.571633 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4glq" event={"ID":"1d9232b8-b9f5-43f9-b31e-eff6c189d91b","Type":"ContainerDied","Data":"32291466750bcbd135477c34808983884cc87a42aaa782b65fd71f3ea4ce3c07"} Dec 04 10:40:44 crc kubenswrapper[4943]: I1204 10:40:44.591754 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4glq" event={"ID":"1d9232b8-b9f5-43f9-b31e-eff6c189d91b","Type":"ContainerStarted","Data":"306514c6a974e978549705d2d9801c6a232170e0f925f5ef8694fab8f24daedd"} Dec 04 10:40:44 crc kubenswrapper[4943]: I1204 10:40:44.609932 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p4glq" podStartSLOduration=3.177081865 podStartE2EDuration="5.609894279s" podCreationTimestamp="2025-12-04 10:40:39 +0000 UTC" firstStartedPulling="2025-12-04 10:40:41.556381761 +0000 UTC m=+1490.145357629" lastFinishedPulling="2025-12-04 10:40:43.989194175 +0000 UTC m=+1492.578170043" observedRunningTime="2025-12-04 10:40:44.607904886 +0000 UTC m=+1493.196880784" watchObservedRunningTime="2025-12-04 10:40:44.609894279 +0000 UTC m=+1493.198870147" Dec 04 10:40:50 crc kubenswrapper[4943]: I1204 10:40:50.329064 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:50 crc kubenswrapper[4943]: I1204 10:40:50.329591 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:50 crc kubenswrapper[4943]: I1204 10:40:50.369986 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:50 crc kubenswrapper[4943]: I1204 10:40:50.735947 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:50 crc kubenswrapper[4943]: I1204 10:40:50.774050 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p4glq"] Dec 04 10:40:52 crc kubenswrapper[4943]: I1204 10:40:52.651585 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p4glq" podUID="1d9232b8-b9f5-43f9-b31e-eff6c189d91b" containerName="registry-server" containerID="cri-o://306514c6a974e978549705d2d9801c6a232170e0f925f5ef8694fab8f24daedd" gracePeriod=2 Dec 04 10:40:53 crc kubenswrapper[4943]: I1204 10:40:53.609806 4943 scope.go:117] "RemoveContainer" containerID="d2390b136890132f4569ff5ff2cb50b6fad4045793a4097fc9cef7eba14c4667" Dec 04 10:40:53 crc kubenswrapper[4943]: I1204 10:40:53.672805 4943 generic.go:334] "Generic (PLEG): container finished" podID="1d9232b8-b9f5-43f9-b31e-eff6c189d91b" containerID="306514c6a974e978549705d2d9801c6a232170e0f925f5ef8694fab8f24daedd" exitCode=0 Dec 04 10:40:53 crc kubenswrapper[4943]: I1204 10:40:53.672905 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4glq" event={"ID":"1d9232b8-b9f5-43f9-b31e-eff6c189d91b","Type":"ContainerDied","Data":"306514c6a974e978549705d2d9801c6a232170e0f925f5ef8694fab8f24daedd"} Dec 04 10:40:53 crc kubenswrapper[4943]: I1204 10:40:53.674834 4943 scope.go:117] "RemoveContainer" containerID="1e42cff7bee1e980d96bb2da3bc885ca5a59eefff74bc4aee1b1033dfea51910" Dec 04 10:40:53 crc kubenswrapper[4943]: I1204 10:40:53.704139 4943 scope.go:117] "RemoveContainer" containerID="eef8d03406a95dc991fb6f7c8dc1584719b2b49f9ded87299333b8d6815c24ac" Dec 04 10:40:53 crc kubenswrapper[4943]: I1204 10:40:53.731576 4943 scope.go:117] "RemoveContainer" containerID="dc54ccd7fcdb9d90906924dc9830ffe2f2a3c3ba6c9237079ceb16bf5006727c" Dec 04 10:40:53 crc kubenswrapper[4943]: I1204 10:40:53.747122 4943 scope.go:117] "RemoveContainer" containerID="2bc153a1c3c30c517d39101699cd07667061e44a498b4d660f6d8f6bc96fee74" Dec 04 10:40:53 crc kubenswrapper[4943]: I1204 10:40:53.780528 4943 scope.go:117] "RemoveContainer" containerID="81d9788be38c3eac800e4924c843ed26d147a2ad0600cfa557d270575a37fa25" Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.191656 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.325603 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-utilities\") pod \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\" (UID: \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\") " Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.325711 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bm5kr\" (UniqueName: \"kubernetes.io/projected/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-kube-api-access-bm5kr\") pod \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\" (UID: \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\") " Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.325770 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-catalog-content\") pod \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\" (UID: \"1d9232b8-b9f5-43f9-b31e-eff6c189d91b\") " Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.326544 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-utilities" (OuterVolumeSpecName: "utilities") pod "1d9232b8-b9f5-43f9-b31e-eff6c189d91b" (UID: "1d9232b8-b9f5-43f9-b31e-eff6c189d91b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.341778 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-kube-api-access-bm5kr" (OuterVolumeSpecName: "kube-api-access-bm5kr") pod "1d9232b8-b9f5-43f9-b31e-eff6c189d91b" (UID: "1d9232b8-b9f5-43f9-b31e-eff6c189d91b"). InnerVolumeSpecName "kube-api-access-bm5kr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.426853 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bm5kr\" (UniqueName: \"kubernetes.io/projected/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-kube-api-access-bm5kr\") on node \"crc\" DevicePath \"\"" Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.426901 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.434440 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d9232b8-b9f5-43f9-b31e-eff6c189d91b" (UID: "1d9232b8-b9f5-43f9-b31e-eff6c189d91b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.528817 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d9232b8-b9f5-43f9-b31e-eff6c189d91b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.689450 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p4glq" event={"ID":"1d9232b8-b9f5-43f9-b31e-eff6c189d91b","Type":"ContainerDied","Data":"fb648bcd221c6ab2921f073cbf3ddad46d18e1f72291f27bca9dc5948a8cbb2e"} Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.689551 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p4glq" Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.689582 4943 scope.go:117] "RemoveContainer" containerID="306514c6a974e978549705d2d9801c6a232170e0f925f5ef8694fab8f24daedd" Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.724882 4943 scope.go:117] "RemoveContainer" containerID="32291466750bcbd135477c34808983884cc87a42aaa782b65fd71f3ea4ce3c07" Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.733013 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p4glq"] Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.745754 4943 scope.go:117] "RemoveContainer" containerID="15d8c3671937450205f3f3c069810ffc189ed3a98e4949675411104bf6452f1d" Dec 04 10:40:55 crc kubenswrapper[4943]: I1204 10:40:55.752347 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p4glq"] Dec 04 10:40:56 crc kubenswrapper[4943]: I1204 10:40:56.572430 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d9232b8-b9f5-43f9-b31e-eff6c189d91b" path="/var/lib/kubelet/pods/1d9232b8-b9f5-43f9-b31e-eff6c189d91b/volumes" Dec 04 10:41:07 crc kubenswrapper[4943]: I1204 10:41:07.800571 4943 generic.go:334] "Generic (PLEG): container finished" podID="53981522-c74a-43ce-be48-271982fe9485" containerID="f4ecac3eb0af694a09914f36ddacf9c5f70401b5e8b41b6a0822cf0946de9c53" exitCode=0 Dec 04 10:41:07 crc kubenswrapper[4943]: I1204 10:41:07.800653 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-hjzgj/must-gather-sscvm" event={"ID":"53981522-c74a-43ce-be48-271982fe9485","Type":"ContainerDied","Data":"f4ecac3eb0af694a09914f36ddacf9c5f70401b5e8b41b6a0822cf0946de9c53"} Dec 04 10:41:07 crc kubenswrapper[4943]: I1204 10:41:07.801936 4943 scope.go:117] "RemoveContainer" containerID="f4ecac3eb0af694a09914f36ddacf9c5f70401b5e8b41b6a0822cf0946de9c53" Dec 04 10:41:08 crc kubenswrapper[4943]: I1204 10:41:08.516882 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hjzgj_must-gather-sscvm_53981522-c74a-43ce-be48-271982fe9485/gather/0.log" Dec 04 10:41:15 crc kubenswrapper[4943]: I1204 10:41:15.516471 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-hjzgj/must-gather-sscvm"] Dec 04 10:41:15 crc kubenswrapper[4943]: I1204 10:41:15.517280 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-hjzgj/must-gather-sscvm" podUID="53981522-c74a-43ce-be48-271982fe9485" containerName="copy" containerID="cri-o://48740a0a3f8cf0d1df14659c143ad2d919bd2050b5dac5b9960ddea795f75a1d" gracePeriod=2 Dec 04 10:41:15 crc kubenswrapper[4943]: I1204 10:41:15.522816 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-hjzgj/must-gather-sscvm"] Dec 04 10:41:15 crc kubenswrapper[4943]: I1204 10:41:15.856426 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hjzgj_must-gather-sscvm_53981522-c74a-43ce-be48-271982fe9485/copy/0.log" Dec 04 10:41:15 crc kubenswrapper[4943]: I1204 10:41:15.856972 4943 generic.go:334] "Generic (PLEG): container finished" podID="53981522-c74a-43ce-be48-271982fe9485" containerID="48740a0a3f8cf0d1df14659c143ad2d919bd2050b5dac5b9960ddea795f75a1d" exitCode=143 Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.432477 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hjzgj_must-gather-sscvm_53981522-c74a-43ce-be48-271982fe9485/copy/0.log" Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.432882 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjzgj/must-gather-sscvm" Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.434108 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q99v\" (UniqueName: \"kubernetes.io/projected/53981522-c74a-43ce-be48-271982fe9485-kube-api-access-6q99v\") pod \"53981522-c74a-43ce-be48-271982fe9485\" (UID: \"53981522-c74a-43ce-be48-271982fe9485\") " Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.434161 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/53981522-c74a-43ce-be48-271982fe9485-must-gather-output\") pod \"53981522-c74a-43ce-be48-271982fe9485\" (UID: \"53981522-c74a-43ce-be48-271982fe9485\") " Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.440571 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53981522-c74a-43ce-be48-271982fe9485-kube-api-access-6q99v" (OuterVolumeSpecName: "kube-api-access-6q99v") pod "53981522-c74a-43ce-be48-271982fe9485" (UID: "53981522-c74a-43ce-be48-271982fe9485"). InnerVolumeSpecName "kube-api-access-6q99v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.500389 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53981522-c74a-43ce-be48-271982fe9485-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "53981522-c74a-43ce-be48-271982fe9485" (UID: "53981522-c74a-43ce-be48-271982fe9485"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.535189 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q99v\" (UniqueName: \"kubernetes.io/projected/53981522-c74a-43ce-be48-271982fe9485-kube-api-access-6q99v\") on node \"crc\" DevicePath \"\"" Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.535249 4943 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/53981522-c74a-43ce-be48-271982fe9485-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.572105 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53981522-c74a-43ce-be48-271982fe9485" path="/var/lib/kubelet/pods/53981522-c74a-43ce-be48-271982fe9485/volumes" Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.864833 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-hjzgj_must-gather-sscvm_53981522-c74a-43ce-be48-271982fe9485/copy/0.log" Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.865546 4943 scope.go:117] "RemoveContainer" containerID="48740a0a3f8cf0d1df14659c143ad2d919bd2050b5dac5b9960ddea795f75a1d" Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.865669 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-hjzgj/must-gather-sscvm" Dec 04 10:41:16 crc kubenswrapper[4943]: I1204 10:41:16.883628 4943 scope.go:117] "RemoveContainer" containerID="f4ecac3eb0af694a09914f36ddacf9c5f70401b5e8b41b6a0822cf0946de9c53" Dec 04 10:41:18 crc kubenswrapper[4943]: I1204 10:41:18.628845 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:41:18 crc kubenswrapper[4943]: I1204 10:41:18.628931 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.832869 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r49ph"] Dec 04 10:41:24 crc kubenswrapper[4943]: E1204 10:41:24.833829 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d9232b8-b9f5-43f9-b31e-eff6c189d91b" containerName="extract-content" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.833862 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d9232b8-b9f5-43f9-b31e-eff6c189d91b" containerName="extract-content" Dec 04 10:41:24 crc kubenswrapper[4943]: E1204 10:41:24.833884 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d9232b8-b9f5-43f9-b31e-eff6c189d91b" containerName="registry-server" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.833893 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d9232b8-b9f5-43f9-b31e-eff6c189d91b" containerName="registry-server" Dec 04 10:41:24 crc kubenswrapper[4943]: E1204 10:41:24.833907 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53981522-c74a-43ce-be48-271982fe9485" containerName="gather" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.833915 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="53981522-c74a-43ce-be48-271982fe9485" containerName="gather" Dec 04 10:41:24 crc kubenswrapper[4943]: E1204 10:41:24.833935 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53981522-c74a-43ce-be48-271982fe9485" containerName="copy" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.833942 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="53981522-c74a-43ce-be48-271982fe9485" containerName="copy" Dec 04 10:41:24 crc kubenswrapper[4943]: E1204 10:41:24.833952 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d9232b8-b9f5-43f9-b31e-eff6c189d91b" containerName="extract-utilities" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.833960 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d9232b8-b9f5-43f9-b31e-eff6c189d91b" containerName="extract-utilities" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.834173 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="53981522-c74a-43ce-be48-271982fe9485" containerName="gather" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.834190 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d9232b8-b9f5-43f9-b31e-eff6c189d91b" containerName="registry-server" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.834219 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="53981522-c74a-43ce-be48-271982fe9485" containerName="copy" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.835369 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.846046 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r49ph"] Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.849642 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-catalog-content\") pod \"certified-operators-r49ph\" (UID: \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\") " pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.849772 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76l9v\" (UniqueName: \"kubernetes.io/projected/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-kube-api-access-76l9v\") pod \"certified-operators-r49ph\" (UID: \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\") " pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.849807 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-utilities\") pod \"certified-operators-r49ph\" (UID: \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\") " pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.950598 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76l9v\" (UniqueName: \"kubernetes.io/projected/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-kube-api-access-76l9v\") pod \"certified-operators-r49ph\" (UID: \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\") " pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.950910 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-utilities\") pod \"certified-operators-r49ph\" (UID: \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\") " pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.951087 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-catalog-content\") pod \"certified-operators-r49ph\" (UID: \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\") " pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.951515 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-utilities\") pod \"certified-operators-r49ph\" (UID: \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\") " pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.951608 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-catalog-content\") pod \"certified-operators-r49ph\" (UID: \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\") " pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:24 crc kubenswrapper[4943]: I1204 10:41:24.970833 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76l9v\" (UniqueName: \"kubernetes.io/projected/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-kube-api-access-76l9v\") pod \"certified-operators-r49ph\" (UID: \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\") " pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:25 crc kubenswrapper[4943]: I1204 10:41:25.155084 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:25 crc kubenswrapper[4943]: I1204 10:41:25.449544 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r49ph"] Dec 04 10:41:25 crc kubenswrapper[4943]: I1204 10:41:25.929537 4943 generic.go:334] "Generic (PLEG): container finished" podID="3c9cdf83-1f86-4c6d-80ad-ea89f30def31" containerID="dd784665a916a6f17dce40f2775b24b346c69f6d97b38f1061afa07199e28596" exitCode=0 Dec 04 10:41:25 crc kubenswrapper[4943]: I1204 10:41:25.929655 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r49ph" event={"ID":"3c9cdf83-1f86-4c6d-80ad-ea89f30def31","Type":"ContainerDied","Data":"dd784665a916a6f17dce40f2775b24b346c69f6d97b38f1061afa07199e28596"} Dec 04 10:41:25 crc kubenswrapper[4943]: I1204 10:41:25.929799 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r49ph" event={"ID":"3c9cdf83-1f86-4c6d-80ad-ea89f30def31","Type":"ContainerStarted","Data":"5e1bffe2b950ec38582a39b458456710ba0e8061d303bdaae333ed9cd2e871f9"} Dec 04 10:41:27 crc kubenswrapper[4943]: I1204 10:41:27.943175 4943 generic.go:334] "Generic (PLEG): container finished" podID="3c9cdf83-1f86-4c6d-80ad-ea89f30def31" containerID="cba1ef33e74e7bbecc706b81690b0e1194027ca46766c7e884ddc005f9cbe358" exitCode=0 Dec 04 10:41:27 crc kubenswrapper[4943]: I1204 10:41:27.943249 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r49ph" event={"ID":"3c9cdf83-1f86-4c6d-80ad-ea89f30def31","Type":"ContainerDied","Data":"cba1ef33e74e7bbecc706b81690b0e1194027ca46766c7e884ddc005f9cbe358"} Dec 04 10:41:28 crc kubenswrapper[4943]: I1204 10:41:28.951060 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r49ph" event={"ID":"3c9cdf83-1f86-4c6d-80ad-ea89f30def31","Type":"ContainerStarted","Data":"4b47417e45ee9e65d8a6abacdbeced443c870d6d355543f429cb6de6d34f9aab"} Dec 04 10:41:28 crc kubenswrapper[4943]: I1204 10:41:28.968224 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r49ph" podStartSLOduration=2.33874759 podStartE2EDuration="4.968163895s" podCreationTimestamp="2025-12-04 10:41:24 +0000 UTC" firstStartedPulling="2025-12-04 10:41:25.932550234 +0000 UTC m=+1534.521526102" lastFinishedPulling="2025-12-04 10:41:28.561966549 +0000 UTC m=+1537.150942407" observedRunningTime="2025-12-04 10:41:28.964842166 +0000 UTC m=+1537.553818054" watchObservedRunningTime="2025-12-04 10:41:28.968163895 +0000 UTC m=+1537.557139763" Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.390048 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sgtv4"] Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.391592 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.405362 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sgtv4"] Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.534159 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67acafd-c638-416a-af2d-61e3b5cf1ef2-utilities\") pod \"community-operators-sgtv4\" (UID: \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\") " pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.534268 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5wlb\" (UniqueName: \"kubernetes.io/projected/c67acafd-c638-416a-af2d-61e3b5cf1ef2-kube-api-access-s5wlb\") pod \"community-operators-sgtv4\" (UID: \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\") " pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.534294 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67acafd-c638-416a-af2d-61e3b5cf1ef2-catalog-content\") pod \"community-operators-sgtv4\" (UID: \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\") " pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.635500 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5wlb\" (UniqueName: \"kubernetes.io/projected/c67acafd-c638-416a-af2d-61e3b5cf1ef2-kube-api-access-s5wlb\") pod \"community-operators-sgtv4\" (UID: \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\") " pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.635807 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67acafd-c638-416a-af2d-61e3b5cf1ef2-catalog-content\") pod \"community-operators-sgtv4\" (UID: \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\") " pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.635924 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67acafd-c638-416a-af2d-61e3b5cf1ef2-utilities\") pod \"community-operators-sgtv4\" (UID: \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\") " pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.636689 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67acafd-c638-416a-af2d-61e3b5cf1ef2-utilities\") pod \"community-operators-sgtv4\" (UID: \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\") " pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.636690 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67acafd-c638-416a-af2d-61e3b5cf1ef2-catalog-content\") pod \"community-operators-sgtv4\" (UID: \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\") " pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.657945 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5wlb\" (UniqueName: \"kubernetes.io/projected/c67acafd-c638-416a-af2d-61e3b5cf1ef2-kube-api-access-s5wlb\") pod \"community-operators-sgtv4\" (UID: \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\") " pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:31 crc kubenswrapper[4943]: I1204 10:41:31.708630 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:32 crc kubenswrapper[4943]: I1204 10:41:32.003154 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sgtv4"] Dec 04 10:41:32 crc kubenswrapper[4943]: I1204 10:41:32.989338 4943 generic.go:334] "Generic (PLEG): container finished" podID="c67acafd-c638-416a-af2d-61e3b5cf1ef2" containerID="246baa44a7d75eca32de20aeab5d46abe579dbe5fa8ef6b82264e1c85d49f09c" exitCode=0 Dec 04 10:41:32 crc kubenswrapper[4943]: I1204 10:41:32.989423 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgtv4" event={"ID":"c67acafd-c638-416a-af2d-61e3b5cf1ef2","Type":"ContainerDied","Data":"246baa44a7d75eca32de20aeab5d46abe579dbe5fa8ef6b82264e1c85d49f09c"} Dec 04 10:41:32 crc kubenswrapper[4943]: I1204 10:41:32.989750 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgtv4" event={"ID":"c67acafd-c638-416a-af2d-61e3b5cf1ef2","Type":"ContainerStarted","Data":"18e0653442d1c22df16386fac20f5a4f3d098383a638b013f99ca567087fe409"} Dec 04 10:41:35 crc kubenswrapper[4943]: I1204 10:41:35.005296 4943 generic.go:334] "Generic (PLEG): container finished" podID="c67acafd-c638-416a-af2d-61e3b5cf1ef2" containerID="d81d2c27f07fdca6bf836cd9095f89bd926e346230110da1d20d5900f24c0ee5" exitCode=0 Dec 04 10:41:35 crc kubenswrapper[4943]: I1204 10:41:35.005406 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgtv4" event={"ID":"c67acafd-c638-416a-af2d-61e3b5cf1ef2","Type":"ContainerDied","Data":"d81d2c27f07fdca6bf836cd9095f89bd926e346230110da1d20d5900f24c0ee5"} Dec 04 10:41:35 crc kubenswrapper[4943]: I1204 10:41:35.155488 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:35 crc kubenswrapper[4943]: I1204 10:41:35.155756 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:35 crc kubenswrapper[4943]: I1204 10:41:35.213902 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:36 crc kubenswrapper[4943]: I1204 10:41:36.021778 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgtv4" event={"ID":"c67acafd-c638-416a-af2d-61e3b5cf1ef2","Type":"ContainerStarted","Data":"58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f"} Dec 04 10:41:36 crc kubenswrapper[4943]: I1204 10:41:36.047536 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sgtv4" podStartSLOduration=2.642554886 podStartE2EDuration="5.047510416s" podCreationTimestamp="2025-12-04 10:41:31 +0000 UTC" firstStartedPulling="2025-12-04 10:41:32.992321453 +0000 UTC m=+1541.581297321" lastFinishedPulling="2025-12-04 10:41:35.397276983 +0000 UTC m=+1543.986252851" observedRunningTime="2025-12-04 10:41:36.046457747 +0000 UTC m=+1544.635433625" watchObservedRunningTime="2025-12-04 10:41:36.047510416 +0000 UTC m=+1544.636486294" Dec 04 10:41:36 crc kubenswrapper[4943]: I1204 10:41:36.068534 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:37 crc kubenswrapper[4943]: I1204 10:41:37.368724 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r49ph"] Dec 04 10:41:39 crc kubenswrapper[4943]: I1204 10:41:39.040631 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r49ph" podUID="3c9cdf83-1f86-4c6d-80ad-ea89f30def31" containerName="registry-server" containerID="cri-o://4b47417e45ee9e65d8a6abacdbeced443c870d6d355543f429cb6de6d34f9aab" gracePeriod=2 Dec 04 10:41:40 crc kubenswrapper[4943]: I1204 10:41:40.053231 4943 generic.go:334] "Generic (PLEG): container finished" podID="3c9cdf83-1f86-4c6d-80ad-ea89f30def31" containerID="4b47417e45ee9e65d8a6abacdbeced443c870d6d355543f429cb6de6d34f9aab" exitCode=0 Dec 04 10:41:40 crc kubenswrapper[4943]: I1204 10:41:40.053245 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r49ph" event={"ID":"3c9cdf83-1f86-4c6d-80ad-ea89f30def31","Type":"ContainerDied","Data":"4b47417e45ee9e65d8a6abacdbeced443c870d6d355543f429cb6de6d34f9aab"} Dec 04 10:41:40 crc kubenswrapper[4943]: I1204 10:41:40.332276 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:40 crc kubenswrapper[4943]: I1204 10:41:40.388038 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76l9v\" (UniqueName: \"kubernetes.io/projected/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-kube-api-access-76l9v\") pod \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\" (UID: \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\") " Dec 04 10:41:40 crc kubenswrapper[4943]: I1204 10:41:40.388141 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-utilities\") pod \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\" (UID: \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\") " Dec 04 10:41:40 crc kubenswrapper[4943]: I1204 10:41:40.388179 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-catalog-content\") pod \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\" (UID: \"3c9cdf83-1f86-4c6d-80ad-ea89f30def31\") " Dec 04 10:41:40 crc kubenswrapper[4943]: I1204 10:41:40.388926 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-utilities" (OuterVolumeSpecName: "utilities") pod "3c9cdf83-1f86-4c6d-80ad-ea89f30def31" (UID: "3c9cdf83-1f86-4c6d-80ad-ea89f30def31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:41:40 crc kubenswrapper[4943]: I1204 10:41:40.394298 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-kube-api-access-76l9v" (OuterVolumeSpecName: "kube-api-access-76l9v") pod "3c9cdf83-1f86-4c6d-80ad-ea89f30def31" (UID: "3c9cdf83-1f86-4c6d-80ad-ea89f30def31"). InnerVolumeSpecName "kube-api-access-76l9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:41:40 crc kubenswrapper[4943]: I1204 10:41:40.434944 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c9cdf83-1f86-4c6d-80ad-ea89f30def31" (UID: "3c9cdf83-1f86-4c6d-80ad-ea89f30def31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:41:40 crc kubenswrapper[4943]: I1204 10:41:40.489692 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76l9v\" (UniqueName: \"kubernetes.io/projected/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-kube-api-access-76l9v\") on node \"crc\" DevicePath \"\"" Dec 04 10:41:40 crc kubenswrapper[4943]: I1204 10:41:40.489730 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:41:40 crc kubenswrapper[4943]: I1204 10:41:40.489739 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c9cdf83-1f86-4c6d-80ad-ea89f30def31-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:41:41 crc kubenswrapper[4943]: I1204 10:41:41.066850 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r49ph" event={"ID":"3c9cdf83-1f86-4c6d-80ad-ea89f30def31","Type":"ContainerDied","Data":"5e1bffe2b950ec38582a39b458456710ba0e8061d303bdaae333ed9cd2e871f9"} Dec 04 10:41:41 crc kubenswrapper[4943]: I1204 10:41:41.066947 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r49ph" Dec 04 10:41:41 crc kubenswrapper[4943]: I1204 10:41:41.067977 4943 scope.go:117] "RemoveContainer" containerID="4b47417e45ee9e65d8a6abacdbeced443c870d6d355543f429cb6de6d34f9aab" Dec 04 10:41:41 crc kubenswrapper[4943]: I1204 10:41:41.095393 4943 scope.go:117] "RemoveContainer" containerID="cba1ef33e74e7bbecc706b81690b0e1194027ca46766c7e884ddc005f9cbe358" Dec 04 10:41:41 crc kubenswrapper[4943]: I1204 10:41:41.098982 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r49ph"] Dec 04 10:41:41 crc kubenswrapper[4943]: I1204 10:41:41.102941 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r49ph"] Dec 04 10:41:41 crc kubenswrapper[4943]: I1204 10:41:41.117212 4943 scope.go:117] "RemoveContainer" containerID="dd784665a916a6f17dce40f2775b24b346c69f6d97b38f1061afa07199e28596" Dec 04 10:41:41 crc kubenswrapper[4943]: I1204 10:41:41.709353 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:41 crc kubenswrapper[4943]: I1204 10:41:41.709434 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:41 crc kubenswrapper[4943]: I1204 10:41:41.780225 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:42 crc kubenswrapper[4943]: I1204 10:41:42.116968 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:42 crc kubenswrapper[4943]: I1204 10:41:42.571978 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c9cdf83-1f86-4c6d-80ad-ea89f30def31" path="/var/lib/kubelet/pods/3c9cdf83-1f86-4c6d-80ad-ea89f30def31/volumes" Dec 04 10:41:43 crc kubenswrapper[4943]: I1204 10:41:43.382585 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sgtv4"] Dec 04 10:41:44 crc kubenswrapper[4943]: I1204 10:41:44.086485 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sgtv4" podUID="c67acafd-c638-416a-af2d-61e3b5cf1ef2" containerName="registry-server" containerID="cri-o://58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f" gracePeriod=2 Dec 04 10:41:44 crc kubenswrapper[4943]: I1204 10:41:44.478055 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:44 crc kubenswrapper[4943]: I1204 10:41:44.561111 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67acafd-c638-416a-af2d-61e3b5cf1ef2-utilities\") pod \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\" (UID: \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\") " Dec 04 10:41:44 crc kubenswrapper[4943]: I1204 10:41:44.561270 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67acafd-c638-416a-af2d-61e3b5cf1ef2-catalog-content\") pod \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\" (UID: \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\") " Dec 04 10:41:44 crc kubenswrapper[4943]: I1204 10:41:44.561331 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5wlb\" (UniqueName: \"kubernetes.io/projected/c67acafd-c638-416a-af2d-61e3b5cf1ef2-kube-api-access-s5wlb\") pod \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\" (UID: \"c67acafd-c638-416a-af2d-61e3b5cf1ef2\") " Dec 04 10:41:44 crc kubenswrapper[4943]: I1204 10:41:44.562787 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c67acafd-c638-416a-af2d-61e3b5cf1ef2-utilities" (OuterVolumeSpecName: "utilities") pod "c67acafd-c638-416a-af2d-61e3b5cf1ef2" (UID: "c67acafd-c638-416a-af2d-61e3b5cf1ef2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:41:44 crc kubenswrapper[4943]: I1204 10:41:44.568254 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c67acafd-c638-416a-af2d-61e3b5cf1ef2-kube-api-access-s5wlb" (OuterVolumeSpecName: "kube-api-access-s5wlb") pod "c67acafd-c638-416a-af2d-61e3b5cf1ef2" (UID: "c67acafd-c638-416a-af2d-61e3b5cf1ef2"). InnerVolumeSpecName "kube-api-access-s5wlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:41:44 crc kubenswrapper[4943]: I1204 10:41:44.662934 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67acafd-c638-416a-af2d-61e3b5cf1ef2-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:41:44 crc kubenswrapper[4943]: I1204 10:41:44.663419 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5wlb\" (UniqueName: \"kubernetes.io/projected/c67acafd-c638-416a-af2d-61e3b5cf1ef2-kube-api-access-s5wlb\") on node \"crc\" DevicePath \"\"" Dec 04 10:41:44 crc kubenswrapper[4943]: I1204 10:41:44.883604 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c67acafd-c638-416a-af2d-61e3b5cf1ef2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c67acafd-c638-416a-af2d-61e3b5cf1ef2" (UID: "c67acafd-c638-416a-af2d-61e3b5cf1ef2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:41:44 crc kubenswrapper[4943]: I1204 10:41:44.967437 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67acafd-c638-416a-af2d-61e3b5cf1ef2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.094106 4943 generic.go:334] "Generic (PLEG): container finished" podID="c67acafd-c638-416a-af2d-61e3b5cf1ef2" containerID="58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f" exitCode=0 Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.094150 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgtv4" event={"ID":"c67acafd-c638-416a-af2d-61e3b5cf1ef2","Type":"ContainerDied","Data":"58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f"} Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.094193 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sgtv4" Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.094239 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sgtv4" event={"ID":"c67acafd-c638-416a-af2d-61e3b5cf1ef2","Type":"ContainerDied","Data":"18e0653442d1c22df16386fac20f5a4f3d098383a638b013f99ca567087fe409"} Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.094268 4943 scope.go:117] "RemoveContainer" containerID="58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f" Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.113951 4943 scope.go:117] "RemoveContainer" containerID="d81d2c27f07fdca6bf836cd9095f89bd926e346230110da1d20d5900f24c0ee5" Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.124285 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sgtv4"] Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.128811 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sgtv4"] Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.150681 4943 scope.go:117] "RemoveContainer" containerID="246baa44a7d75eca32de20aeab5d46abe579dbe5fa8ef6b82264e1c85d49f09c" Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.165685 4943 scope.go:117] "RemoveContainer" containerID="58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f" Dec 04 10:41:45 crc kubenswrapper[4943]: E1204 10:41:45.166478 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f\": container with ID starting with 58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f not found: ID does not exist" containerID="58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f" Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.166580 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f"} err="failed to get container status \"58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f\": rpc error: code = NotFound desc = could not find container \"58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f\": container with ID starting with 58b4abd4e6f6d919d6b0d017e5c9590b50c36797df4cec722d3a0fe6e062322f not found: ID does not exist" Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.166671 4943 scope.go:117] "RemoveContainer" containerID="d81d2c27f07fdca6bf836cd9095f89bd926e346230110da1d20d5900f24c0ee5" Dec 04 10:41:45 crc kubenswrapper[4943]: E1204 10:41:45.167120 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d81d2c27f07fdca6bf836cd9095f89bd926e346230110da1d20d5900f24c0ee5\": container with ID starting with d81d2c27f07fdca6bf836cd9095f89bd926e346230110da1d20d5900f24c0ee5 not found: ID does not exist" containerID="d81d2c27f07fdca6bf836cd9095f89bd926e346230110da1d20d5900f24c0ee5" Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.167152 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d81d2c27f07fdca6bf836cd9095f89bd926e346230110da1d20d5900f24c0ee5"} err="failed to get container status \"d81d2c27f07fdca6bf836cd9095f89bd926e346230110da1d20d5900f24c0ee5\": rpc error: code = NotFound desc = could not find container \"d81d2c27f07fdca6bf836cd9095f89bd926e346230110da1d20d5900f24c0ee5\": container with ID starting with d81d2c27f07fdca6bf836cd9095f89bd926e346230110da1d20d5900f24c0ee5 not found: ID does not exist" Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.167174 4943 scope.go:117] "RemoveContainer" containerID="246baa44a7d75eca32de20aeab5d46abe579dbe5fa8ef6b82264e1c85d49f09c" Dec 04 10:41:45 crc kubenswrapper[4943]: E1204 10:41:45.167482 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"246baa44a7d75eca32de20aeab5d46abe579dbe5fa8ef6b82264e1c85d49f09c\": container with ID starting with 246baa44a7d75eca32de20aeab5d46abe579dbe5fa8ef6b82264e1c85d49f09c not found: ID does not exist" containerID="246baa44a7d75eca32de20aeab5d46abe579dbe5fa8ef6b82264e1c85d49f09c" Dec 04 10:41:45 crc kubenswrapper[4943]: I1204 10:41:45.167533 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"246baa44a7d75eca32de20aeab5d46abe579dbe5fa8ef6b82264e1c85d49f09c"} err="failed to get container status \"246baa44a7d75eca32de20aeab5d46abe579dbe5fa8ef6b82264e1c85d49f09c\": rpc error: code = NotFound desc = could not find container \"246baa44a7d75eca32de20aeab5d46abe579dbe5fa8ef6b82264e1c85d49f09c\": container with ID starting with 246baa44a7d75eca32de20aeab5d46abe579dbe5fa8ef6b82264e1c85d49f09c not found: ID does not exist" Dec 04 10:41:46 crc kubenswrapper[4943]: I1204 10:41:46.572318 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c67acafd-c638-416a-af2d-61e3b5cf1ef2" path="/var/lib/kubelet/pods/c67acafd-c638-416a-af2d-61e3b5cf1ef2/volumes" Dec 04 10:41:48 crc kubenswrapper[4943]: I1204 10:41:48.629279 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:41:48 crc kubenswrapper[4943]: I1204 10:41:48.629698 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:41:53 crc kubenswrapper[4943]: I1204 10:41:53.840360 4943 scope.go:117] "RemoveContainer" containerID="b5b9f96b5feb0d6fb0d9f40cc14b1870320f8fefa81b6b1042864b78b72a1282" Dec 04 10:41:53 crc kubenswrapper[4943]: I1204 10:41:53.880187 4943 scope.go:117] "RemoveContainer" containerID="8d59862a25addea490182e432a477a56450b82e0ce71280a3fa7c5e66542d704" Dec 04 10:41:53 crc kubenswrapper[4943]: I1204 10:41:53.892779 4943 scope.go:117] "RemoveContainer" containerID="97d1e4bf9571948da1166e8ff207a37343c355e16f45d3d9629e3c6d1ee16256" Dec 04 10:41:53 crc kubenswrapper[4943]: I1204 10:41:53.911768 4943 scope.go:117] "RemoveContainer" containerID="52589586ed94a0549a6570004e167d9ad64bca476673e1655a83ec0bcdefe3a4" Dec 04 10:41:53 crc kubenswrapper[4943]: I1204 10:41:53.936586 4943 scope.go:117] "RemoveContainer" containerID="1e953ac8ed499eabcf2d34c691af5981551178fdb696ff7cba69f38327c886b1" Dec 04 10:41:53 crc kubenswrapper[4943]: I1204 10:41:53.963321 4943 scope.go:117] "RemoveContainer" containerID="a69e6ccf43199cde628626de373ebac1c8bf085abb7dadd544d0c92c218a6b90" Dec 04 10:41:53 crc kubenswrapper[4943]: I1204 10:41:53.975656 4943 scope.go:117] "RemoveContainer" containerID="272ee3e9542601fd57eaa015217a3befda4353aacea9cb30765e43e04c4d9ceb" Dec 04 10:42:18 crc kubenswrapper[4943]: I1204 10:42:18.629448 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:42:18 crc kubenswrapper[4943]: I1204 10:42:18.629903 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:42:18 crc kubenswrapper[4943]: I1204 10:42:18.629951 4943 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" Dec 04 10:42:18 crc kubenswrapper[4943]: I1204 10:42:18.630532 4943 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520"} pod="openshift-machine-config-operator/machine-config-daemon-kswzd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 10:42:18 crc kubenswrapper[4943]: I1204 10:42:18.630627 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" containerID="cri-o://a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" gracePeriod=600 Dec 04 10:42:18 crc kubenswrapper[4943]: E1204 10:42:18.760570 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:42:19 crc kubenswrapper[4943]: I1204 10:42:19.295631 4943 generic.go:334] "Generic (PLEG): container finished" podID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" exitCode=0 Dec 04 10:42:19 crc kubenswrapper[4943]: I1204 10:42:19.295714 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerDied","Data":"a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520"} Dec 04 10:42:19 crc kubenswrapper[4943]: I1204 10:42:19.295980 4943 scope.go:117] "RemoveContainer" containerID="7e58a73db13bfe564778718e69895f4217bb13940aaf3f0a781ae48e378dfd34" Dec 04 10:42:19 crc kubenswrapper[4943]: I1204 10:42:19.296673 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:42:19 crc kubenswrapper[4943]: E1204 10:42:19.297037 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:42:34 crc kubenswrapper[4943]: I1204 10:42:34.564832 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:42:34 crc kubenswrapper[4943]: E1204 10:42:34.565745 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:42:47 crc kubenswrapper[4943]: I1204 10:42:47.564953 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:42:47 crc kubenswrapper[4943]: E1204 10:42:47.565685 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.646300 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vlm2g"] Dec 04 10:42:49 crc kubenswrapper[4943]: E1204 10:42:49.646806 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67acafd-c638-416a-af2d-61e3b5cf1ef2" containerName="registry-server" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.646898 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67acafd-c638-416a-af2d-61e3b5cf1ef2" containerName="registry-server" Dec 04 10:42:49 crc kubenswrapper[4943]: E1204 10:42:49.646925 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c9cdf83-1f86-4c6d-80ad-ea89f30def31" containerName="extract-content" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.646934 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c9cdf83-1f86-4c6d-80ad-ea89f30def31" containerName="extract-content" Dec 04 10:42:49 crc kubenswrapper[4943]: E1204 10:42:49.646952 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67acafd-c638-416a-af2d-61e3b5cf1ef2" containerName="extract-utilities" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.646961 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67acafd-c638-416a-af2d-61e3b5cf1ef2" containerName="extract-utilities" Dec 04 10:42:49 crc kubenswrapper[4943]: E1204 10:42:49.646981 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67acafd-c638-416a-af2d-61e3b5cf1ef2" containerName="extract-content" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.646990 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67acafd-c638-416a-af2d-61e3b5cf1ef2" containerName="extract-content" Dec 04 10:42:49 crc kubenswrapper[4943]: E1204 10:42:49.647007 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c9cdf83-1f86-4c6d-80ad-ea89f30def31" containerName="extract-utilities" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.647015 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c9cdf83-1f86-4c6d-80ad-ea89f30def31" containerName="extract-utilities" Dec 04 10:42:49 crc kubenswrapper[4943]: E1204 10:42:49.647027 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c9cdf83-1f86-4c6d-80ad-ea89f30def31" containerName="registry-server" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.647036 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c9cdf83-1f86-4c6d-80ad-ea89f30def31" containerName="registry-server" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.647252 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="c67acafd-c638-416a-af2d-61e3b5cf1ef2" containerName="registry-server" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.647267 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c9cdf83-1f86-4c6d-80ad-ea89f30def31" containerName="registry-server" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.650906 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.654021 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vlm2g"] Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.665983 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvf5v\" (UniqueName: \"kubernetes.io/projected/c23644e5-02d3-4354-b070-8acd3f86cf5a-kube-api-access-fvf5v\") pod \"redhat-marketplace-vlm2g\" (UID: \"c23644e5-02d3-4354-b070-8acd3f86cf5a\") " pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.666336 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c23644e5-02d3-4354-b070-8acd3f86cf5a-utilities\") pod \"redhat-marketplace-vlm2g\" (UID: \"c23644e5-02d3-4354-b070-8acd3f86cf5a\") " pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.666510 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c23644e5-02d3-4354-b070-8acd3f86cf5a-catalog-content\") pod \"redhat-marketplace-vlm2g\" (UID: \"c23644e5-02d3-4354-b070-8acd3f86cf5a\") " pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.767965 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c23644e5-02d3-4354-b070-8acd3f86cf5a-catalog-content\") pod \"redhat-marketplace-vlm2g\" (UID: \"c23644e5-02d3-4354-b070-8acd3f86cf5a\") " pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.768050 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvf5v\" (UniqueName: \"kubernetes.io/projected/c23644e5-02d3-4354-b070-8acd3f86cf5a-kube-api-access-fvf5v\") pod \"redhat-marketplace-vlm2g\" (UID: \"c23644e5-02d3-4354-b070-8acd3f86cf5a\") " pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.768074 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c23644e5-02d3-4354-b070-8acd3f86cf5a-utilities\") pod \"redhat-marketplace-vlm2g\" (UID: \"c23644e5-02d3-4354-b070-8acd3f86cf5a\") " pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.768559 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c23644e5-02d3-4354-b070-8acd3f86cf5a-catalog-content\") pod \"redhat-marketplace-vlm2g\" (UID: \"c23644e5-02d3-4354-b070-8acd3f86cf5a\") " pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.768591 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c23644e5-02d3-4354-b070-8acd3f86cf5a-utilities\") pod \"redhat-marketplace-vlm2g\" (UID: \"c23644e5-02d3-4354-b070-8acd3f86cf5a\") " pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.790082 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvf5v\" (UniqueName: \"kubernetes.io/projected/c23644e5-02d3-4354-b070-8acd3f86cf5a-kube-api-access-fvf5v\") pod \"redhat-marketplace-vlm2g\" (UID: \"c23644e5-02d3-4354-b070-8acd3f86cf5a\") " pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:42:49 crc kubenswrapper[4943]: I1204 10:42:49.970527 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:42:50 crc kubenswrapper[4943]: I1204 10:42:50.177846 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vlm2g"] Dec 04 10:42:50 crc kubenswrapper[4943]: I1204 10:42:50.508797 4943 generic.go:334] "Generic (PLEG): container finished" podID="c23644e5-02d3-4354-b070-8acd3f86cf5a" containerID="61f56d1aa7c3d2f4d61f0a1ae4d2c132fbbb2c6b2fabeac64fdde4e8a437de24" exitCode=0 Dec 04 10:42:50 crc kubenswrapper[4943]: I1204 10:42:50.508909 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vlm2g" event={"ID":"c23644e5-02d3-4354-b070-8acd3f86cf5a","Type":"ContainerDied","Data":"61f56d1aa7c3d2f4d61f0a1ae4d2c132fbbb2c6b2fabeac64fdde4e8a437de24"} Dec 04 10:42:50 crc kubenswrapper[4943]: I1204 10:42:50.508968 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vlm2g" event={"ID":"c23644e5-02d3-4354-b070-8acd3f86cf5a","Type":"ContainerStarted","Data":"47697fbab54a7375131e07478b7070dab249d89c275dcba4750b1c85749951e1"} Dec 04 10:42:51 crc kubenswrapper[4943]: I1204 10:42:51.515720 4943 generic.go:334] "Generic (PLEG): container finished" podID="c23644e5-02d3-4354-b070-8acd3f86cf5a" containerID="085229b473b4f502a5843a9e8a8a6634eb17965f1f389f096d0b5f9eeca8ffd2" exitCode=0 Dec 04 10:42:51 crc kubenswrapper[4943]: I1204 10:42:51.515821 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vlm2g" event={"ID":"c23644e5-02d3-4354-b070-8acd3f86cf5a","Type":"ContainerDied","Data":"085229b473b4f502a5843a9e8a8a6634eb17965f1f389f096d0b5f9eeca8ffd2"} Dec 04 10:42:52 crc kubenswrapper[4943]: I1204 10:42:52.526341 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vlm2g" event={"ID":"c23644e5-02d3-4354-b070-8acd3f86cf5a","Type":"ContainerStarted","Data":"ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf"} Dec 04 10:42:52 crc kubenswrapper[4943]: I1204 10:42:52.548072 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vlm2g" podStartSLOduration=2.030854352 podStartE2EDuration="3.548035586s" podCreationTimestamp="2025-12-04 10:42:49 +0000 UTC" firstStartedPulling="2025-12-04 10:42:50.510358767 +0000 UTC m=+1619.099334645" lastFinishedPulling="2025-12-04 10:42:52.027540011 +0000 UTC m=+1620.616515879" observedRunningTime="2025-12-04 10:42:52.544862938 +0000 UTC m=+1621.133838826" watchObservedRunningTime="2025-12-04 10:42:52.548035586 +0000 UTC m=+1621.137011454" Dec 04 10:42:59 crc kubenswrapper[4943]: I1204 10:42:59.971526 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:42:59 crc kubenswrapper[4943]: I1204 10:42:59.972252 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:43:00 crc kubenswrapper[4943]: I1204 10:43:00.063583 4943 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:43:00 crc kubenswrapper[4943]: I1204 10:43:00.603074 4943 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:43:00 crc kubenswrapper[4943]: I1204 10:43:00.644848 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vlm2g"] Dec 04 10:43:02 crc kubenswrapper[4943]: I1204 10:43:02.571356 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:43:02 crc kubenswrapper[4943]: E1204 10:43:02.571717 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:43:02 crc kubenswrapper[4943]: I1204 10:43:02.581403 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vlm2g" podUID="c23644e5-02d3-4354-b070-8acd3f86cf5a" containerName="registry-server" containerID="cri-o://ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf" gracePeriod=2 Dec 04 10:43:02 crc kubenswrapper[4943]: I1204 10:43:02.916813 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:43:02 crc kubenswrapper[4943]: I1204 10:43:02.946256 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c23644e5-02d3-4354-b070-8acd3f86cf5a-catalog-content\") pod \"c23644e5-02d3-4354-b070-8acd3f86cf5a\" (UID: \"c23644e5-02d3-4354-b070-8acd3f86cf5a\") " Dec 04 10:43:02 crc kubenswrapper[4943]: I1204 10:43:02.946346 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvf5v\" (UniqueName: \"kubernetes.io/projected/c23644e5-02d3-4354-b070-8acd3f86cf5a-kube-api-access-fvf5v\") pod \"c23644e5-02d3-4354-b070-8acd3f86cf5a\" (UID: \"c23644e5-02d3-4354-b070-8acd3f86cf5a\") " Dec 04 10:43:02 crc kubenswrapper[4943]: I1204 10:43:02.946424 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c23644e5-02d3-4354-b070-8acd3f86cf5a-utilities\") pod \"c23644e5-02d3-4354-b070-8acd3f86cf5a\" (UID: \"c23644e5-02d3-4354-b070-8acd3f86cf5a\") " Dec 04 10:43:02 crc kubenswrapper[4943]: I1204 10:43:02.947841 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c23644e5-02d3-4354-b070-8acd3f86cf5a-utilities" (OuterVolumeSpecName: "utilities") pod "c23644e5-02d3-4354-b070-8acd3f86cf5a" (UID: "c23644e5-02d3-4354-b070-8acd3f86cf5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:43:02 crc kubenswrapper[4943]: I1204 10:43:02.954402 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c23644e5-02d3-4354-b070-8acd3f86cf5a-kube-api-access-fvf5v" (OuterVolumeSpecName: "kube-api-access-fvf5v") pod "c23644e5-02d3-4354-b070-8acd3f86cf5a" (UID: "c23644e5-02d3-4354-b070-8acd3f86cf5a"). InnerVolumeSpecName "kube-api-access-fvf5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:43:02 crc kubenswrapper[4943]: I1204 10:43:02.967016 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c23644e5-02d3-4354-b070-8acd3f86cf5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c23644e5-02d3-4354-b070-8acd3f86cf5a" (UID: "c23644e5-02d3-4354-b070-8acd3f86cf5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.047447 4943 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c23644e5-02d3-4354-b070-8acd3f86cf5a-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.047483 4943 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c23644e5-02d3-4354-b070-8acd3f86cf5a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.047494 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvf5v\" (UniqueName: \"kubernetes.io/projected/c23644e5-02d3-4354-b070-8acd3f86cf5a-kube-api-access-fvf5v\") on node \"crc\" DevicePath \"\"" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.591344 4943 generic.go:334] "Generic (PLEG): container finished" podID="c23644e5-02d3-4354-b070-8acd3f86cf5a" containerID="ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf" exitCode=0 Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.591400 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vlm2g" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.591395 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vlm2g" event={"ID":"c23644e5-02d3-4354-b070-8acd3f86cf5a","Type":"ContainerDied","Data":"ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf"} Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.591571 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vlm2g" event={"ID":"c23644e5-02d3-4354-b070-8acd3f86cf5a","Type":"ContainerDied","Data":"47697fbab54a7375131e07478b7070dab249d89c275dcba4750b1c85749951e1"} Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.591666 4943 scope.go:117] "RemoveContainer" containerID="ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.611933 4943 scope.go:117] "RemoveContainer" containerID="085229b473b4f502a5843a9e8a8a6634eb17965f1f389f096d0b5f9eeca8ffd2" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.624720 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vlm2g"] Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.629935 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vlm2g"] Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.634058 4943 scope.go:117] "RemoveContainer" containerID="61f56d1aa7c3d2f4d61f0a1ae4d2c132fbbb2c6b2fabeac64fdde4e8a437de24" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.652495 4943 scope.go:117] "RemoveContainer" containerID="ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf" Dec 04 10:43:03 crc kubenswrapper[4943]: E1204 10:43:03.652944 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf\": container with ID starting with ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf not found: ID does not exist" containerID="ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.653073 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf"} err="failed to get container status \"ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf\": rpc error: code = NotFound desc = could not find container \"ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf\": container with ID starting with ab437bbb3ce1d6823b01710ca2986f99f95b39c2782af7d987008eb0a3bdd3bf not found: ID does not exist" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.653159 4943 scope.go:117] "RemoveContainer" containerID="085229b473b4f502a5843a9e8a8a6634eb17965f1f389f096d0b5f9eeca8ffd2" Dec 04 10:43:03 crc kubenswrapper[4943]: E1204 10:43:03.653769 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"085229b473b4f502a5843a9e8a8a6634eb17965f1f389f096d0b5f9eeca8ffd2\": container with ID starting with 085229b473b4f502a5843a9e8a8a6634eb17965f1f389f096d0b5f9eeca8ffd2 not found: ID does not exist" containerID="085229b473b4f502a5843a9e8a8a6634eb17965f1f389f096d0b5f9eeca8ffd2" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.653807 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"085229b473b4f502a5843a9e8a8a6634eb17965f1f389f096d0b5f9eeca8ffd2"} err="failed to get container status \"085229b473b4f502a5843a9e8a8a6634eb17965f1f389f096d0b5f9eeca8ffd2\": rpc error: code = NotFound desc = could not find container \"085229b473b4f502a5843a9e8a8a6634eb17965f1f389f096d0b5f9eeca8ffd2\": container with ID starting with 085229b473b4f502a5843a9e8a8a6634eb17965f1f389f096d0b5f9eeca8ffd2 not found: ID does not exist" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.653834 4943 scope.go:117] "RemoveContainer" containerID="61f56d1aa7c3d2f4d61f0a1ae4d2c132fbbb2c6b2fabeac64fdde4e8a437de24" Dec 04 10:43:03 crc kubenswrapper[4943]: E1204 10:43:03.654103 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61f56d1aa7c3d2f4d61f0a1ae4d2c132fbbb2c6b2fabeac64fdde4e8a437de24\": container with ID starting with 61f56d1aa7c3d2f4d61f0a1ae4d2c132fbbb2c6b2fabeac64fdde4e8a437de24 not found: ID does not exist" containerID="61f56d1aa7c3d2f4d61f0a1ae4d2c132fbbb2c6b2fabeac64fdde4e8a437de24" Dec 04 10:43:03 crc kubenswrapper[4943]: I1204 10:43:03.654131 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61f56d1aa7c3d2f4d61f0a1ae4d2c132fbbb2c6b2fabeac64fdde4e8a437de24"} err="failed to get container status \"61f56d1aa7c3d2f4d61f0a1ae4d2c132fbbb2c6b2fabeac64fdde4e8a437de24\": rpc error: code = NotFound desc = could not find container \"61f56d1aa7c3d2f4d61f0a1ae4d2c132fbbb2c6b2fabeac64fdde4e8a437de24\": container with ID starting with 61f56d1aa7c3d2f4d61f0a1ae4d2c132fbbb2c6b2fabeac64fdde4e8a437de24 not found: ID does not exist" Dec 04 10:43:04 crc kubenswrapper[4943]: I1204 10:43:04.572244 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c23644e5-02d3-4354-b070-8acd3f86cf5a" path="/var/lib/kubelet/pods/c23644e5-02d3-4354-b070-8acd3f86cf5a/volumes" Dec 04 10:43:14 crc kubenswrapper[4943]: I1204 10:43:14.564733 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:43:14 crc kubenswrapper[4943]: E1204 10:43:14.565696 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:43:27 crc kubenswrapper[4943]: I1204 10:43:27.564322 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:43:27 crc kubenswrapper[4943]: E1204 10:43:27.565045 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:43:38 crc kubenswrapper[4943]: I1204 10:43:38.565233 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:43:38 crc kubenswrapper[4943]: E1204 10:43:38.567308 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:43:48 crc kubenswrapper[4943]: I1204 10:43:48.818123 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-ttvvb/must-gather-nvch6"] Dec 04 10:43:48 crc kubenswrapper[4943]: E1204 10:43:48.818995 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c23644e5-02d3-4354-b070-8acd3f86cf5a" containerName="extract-content" Dec 04 10:43:48 crc kubenswrapper[4943]: I1204 10:43:48.819012 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="c23644e5-02d3-4354-b070-8acd3f86cf5a" containerName="extract-content" Dec 04 10:43:48 crc kubenswrapper[4943]: E1204 10:43:48.819027 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c23644e5-02d3-4354-b070-8acd3f86cf5a" containerName="extract-utilities" Dec 04 10:43:48 crc kubenswrapper[4943]: I1204 10:43:48.819036 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="c23644e5-02d3-4354-b070-8acd3f86cf5a" containerName="extract-utilities" Dec 04 10:43:48 crc kubenswrapper[4943]: E1204 10:43:48.819062 4943 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c23644e5-02d3-4354-b070-8acd3f86cf5a" containerName="registry-server" Dec 04 10:43:48 crc kubenswrapper[4943]: I1204 10:43:48.819070 4943 state_mem.go:107] "Deleted CPUSet assignment" podUID="c23644e5-02d3-4354-b070-8acd3f86cf5a" containerName="registry-server" Dec 04 10:43:48 crc kubenswrapper[4943]: I1204 10:43:48.819241 4943 memory_manager.go:354] "RemoveStaleState removing state" podUID="c23644e5-02d3-4354-b070-8acd3f86cf5a" containerName="registry-server" Dec 04 10:43:48 crc kubenswrapper[4943]: I1204 10:43:48.819977 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttvvb/must-gather-nvch6" Dec 04 10:43:48 crc kubenswrapper[4943]: I1204 10:43:48.827819 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ttvvb"/"kube-root-ca.crt" Dec 04 10:43:48 crc kubenswrapper[4943]: I1204 10:43:48.828113 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-ttvvb"/"openshift-service-ca.crt" Dec 04 10:43:48 crc kubenswrapper[4943]: I1204 10:43:48.860736 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ttvvb/must-gather-nvch6"] Dec 04 10:43:48 crc kubenswrapper[4943]: I1204 10:43:48.904952 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcds8\" (UniqueName: \"kubernetes.io/projected/32ef8b0a-14c6-47b8-8052-d007e98c3e6c-kube-api-access-hcds8\") pod \"must-gather-nvch6\" (UID: \"32ef8b0a-14c6-47b8-8052-d007e98c3e6c\") " pod="openshift-must-gather-ttvvb/must-gather-nvch6" Dec 04 10:43:48 crc kubenswrapper[4943]: I1204 10:43:48.905085 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/32ef8b0a-14c6-47b8-8052-d007e98c3e6c-must-gather-output\") pod \"must-gather-nvch6\" (UID: \"32ef8b0a-14c6-47b8-8052-d007e98c3e6c\") " pod="openshift-must-gather-ttvvb/must-gather-nvch6" Dec 04 10:43:49 crc kubenswrapper[4943]: I1204 10:43:49.006127 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/32ef8b0a-14c6-47b8-8052-d007e98c3e6c-must-gather-output\") pod \"must-gather-nvch6\" (UID: \"32ef8b0a-14c6-47b8-8052-d007e98c3e6c\") " pod="openshift-must-gather-ttvvb/must-gather-nvch6" Dec 04 10:43:49 crc kubenswrapper[4943]: I1204 10:43:49.006264 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcds8\" (UniqueName: \"kubernetes.io/projected/32ef8b0a-14c6-47b8-8052-d007e98c3e6c-kube-api-access-hcds8\") pod \"must-gather-nvch6\" (UID: \"32ef8b0a-14c6-47b8-8052-d007e98c3e6c\") " pod="openshift-must-gather-ttvvb/must-gather-nvch6" Dec 04 10:43:49 crc kubenswrapper[4943]: I1204 10:43:49.007072 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/32ef8b0a-14c6-47b8-8052-d007e98c3e6c-must-gather-output\") pod \"must-gather-nvch6\" (UID: \"32ef8b0a-14c6-47b8-8052-d007e98c3e6c\") " pod="openshift-must-gather-ttvvb/must-gather-nvch6" Dec 04 10:43:49 crc kubenswrapper[4943]: I1204 10:43:49.025418 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcds8\" (UniqueName: \"kubernetes.io/projected/32ef8b0a-14c6-47b8-8052-d007e98c3e6c-kube-api-access-hcds8\") pod \"must-gather-nvch6\" (UID: \"32ef8b0a-14c6-47b8-8052-d007e98c3e6c\") " pod="openshift-must-gather-ttvvb/must-gather-nvch6" Dec 04 10:43:49 crc kubenswrapper[4943]: I1204 10:43:49.139240 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttvvb/must-gather-nvch6" Dec 04 10:43:49 crc kubenswrapper[4943]: I1204 10:43:49.339183 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-ttvvb/must-gather-nvch6"] Dec 04 10:43:49 crc kubenswrapper[4943]: I1204 10:43:49.883839 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttvvb/must-gather-nvch6" event={"ID":"32ef8b0a-14c6-47b8-8052-d007e98c3e6c","Type":"ContainerStarted","Data":"74d8c0aa9b418ef18dcc5c7e7ac5a6bc5d136ed19f14e01bf4acad861846f190"} Dec 04 10:43:50 crc kubenswrapper[4943]: I1204 10:43:50.892285 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttvvb/must-gather-nvch6" event={"ID":"32ef8b0a-14c6-47b8-8052-d007e98c3e6c","Type":"ContainerStarted","Data":"56653d07a62fd3846df3301a93be11f647348b2e6623221a21431ae19f39a1ad"} Dec 04 10:43:50 crc kubenswrapper[4943]: I1204 10:43:50.892638 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttvvb/must-gather-nvch6" event={"ID":"32ef8b0a-14c6-47b8-8052-d007e98c3e6c","Type":"ContainerStarted","Data":"b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b"} Dec 04 10:43:50 crc kubenswrapper[4943]: I1204 10:43:50.907511 4943 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-ttvvb/must-gather-nvch6" podStartSLOduration=2.907492891 podStartE2EDuration="2.907492891s" podCreationTimestamp="2025-12-04 10:43:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 10:43:50.905860646 +0000 UTC m=+1679.494836514" watchObservedRunningTime="2025-12-04 10:43:50.907492891 +0000 UTC m=+1679.496468759" Dec 04 10:43:52 crc kubenswrapper[4943]: I1204 10:43:52.567819 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:43:52 crc kubenswrapper[4943]: E1204 10:43:52.568054 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:43:54 crc kubenswrapper[4943]: I1204 10:43:54.106531 4943 scope.go:117] "RemoveContainer" containerID="9716c02c743af52efd8e9b59789dfbb5f003a14d06f53dcf23af6ca00f6ad0ec" Dec 04 10:43:54 crc kubenswrapper[4943]: I1204 10:43:54.159293 4943 scope.go:117] "RemoveContainer" containerID="af56e0e6966666528f27742b897bea116befeb238c729c0b5ba875f4ecffea46" Dec 04 10:43:54 crc kubenswrapper[4943]: I1204 10:43:54.175718 4943 scope.go:117] "RemoveContainer" containerID="6066e7ec929e202bc35aa6fe00dabaa4bb894b0e31f40d0151ca84aed43b0171" Dec 04 10:43:54 crc kubenswrapper[4943]: I1204 10:43:54.195678 4943 scope.go:117] "RemoveContainer" containerID="1f4ea23318021a79c31454ef74765c616fe201b11bff0faac5fcea5243e56dcc" Dec 04 10:44:03 crc kubenswrapper[4943]: I1204 10:44:03.565112 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:44:03 crc kubenswrapper[4943]: E1204 10:44:03.565882 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:44:16 crc kubenswrapper[4943]: I1204 10:44:16.568487 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:44:16 crc kubenswrapper[4943]: E1204 10:44:16.569592 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:44:31 crc kubenswrapper[4943]: I1204 10:44:31.565183 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:44:31 crc kubenswrapper[4943]: E1204 10:44:31.565829 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:44:32 crc kubenswrapper[4943]: I1204 10:44:32.352794 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-vhxkd_5ef5c158-a074-47df-a4bb-2e1cab760745/control-plane-machine-set-operator/0.log" Dec 04 10:44:32 crc kubenswrapper[4943]: I1204 10:44:32.488167 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8sdsb_1f2720fe-737e-4562-9fc1-d0e52e0d99a2/kube-rbac-proxy/0.log" Dec 04 10:44:32 crc kubenswrapper[4943]: I1204 10:44:32.519149 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-8sdsb_1f2720fe-737e-4562-9fc1-d0e52e0d99a2/machine-api-operator/0.log" Dec 04 10:44:46 crc kubenswrapper[4943]: I1204 10:44:46.564723 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:44:46 crc kubenswrapper[4943]: E1204 10:44:46.565592 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:44:46 crc kubenswrapper[4943]: I1204 10:44:46.802142 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-hm4b7_71342ec6-c2cb-4277-a567-e13aabab1db3/controller/0.log" Dec 04 10:44:46 crc kubenswrapper[4943]: I1204 10:44:46.810890 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-hm4b7_71342ec6-c2cb-4277-a567-e13aabab1db3/kube-rbac-proxy/0.log" Dec 04 10:44:46 crc kubenswrapper[4943]: I1204 10:44:46.956502 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-frr-files/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.143238 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-frr-files/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.148739 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-reloader/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.153674 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-metrics/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.178541 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-reloader/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.418056 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-metrics/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.449256 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-metrics/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.461709 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-reloader/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.488025 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-frr-files/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.690319 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/controller/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.692668 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-frr-files/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.704903 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-reloader/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.718065 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/cp-metrics/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.907183 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/kube-rbac-proxy/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.935356 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/frr-metrics/0.log" Dec 04 10:44:47 crc kubenswrapper[4943]: I1204 10:44:47.938475 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/kube-rbac-proxy-frr/0.log" Dec 04 10:44:48 crc kubenswrapper[4943]: I1204 10:44:48.078366 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/reloader/0.log" Dec 04 10:44:48 crc kubenswrapper[4943]: I1204 10:44:48.134568 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-t89n4_6104b8e7-773f-4873-912f-5269972dfac9/frr-k8s-webhook-server/0.log" Dec 04 10:44:48 crc kubenswrapper[4943]: I1204 10:44:48.231739 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-k2g9w_07356b24-d6a8-4b4f-8d78-b73a19e05e7d/frr/0.log" Dec 04 10:44:48 crc kubenswrapper[4943]: I1204 10:44:48.344961 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5697dbbbd-r8zls_0ca89ccd-b04f-480d-9a4e-a96b534b3ca6/manager/0.log" Dec 04 10:44:48 crc kubenswrapper[4943]: I1204 10:44:48.453617 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-57ffff8ff7-k42kt_4dfd2ad5-8ff8-4d1e-8976-e1cbc87d1716/webhook-server/0.log" Dec 04 10:44:48 crc kubenswrapper[4943]: I1204 10:44:48.535304 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sd2np_b473a6fb-38de-4766-b414-56f3c0496ff6/kube-rbac-proxy/0.log" Dec 04 10:44:48 crc kubenswrapper[4943]: I1204 10:44:48.623942 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sd2np_b473a6fb-38de-4766-b414-56f3c0496ff6/speaker/0.log" Dec 04 10:44:54 crc kubenswrapper[4943]: I1204 10:44:54.297880 4943 scope.go:117] "RemoveContainer" containerID="2851d6f7b6ae0d36fedbeb621349219674f4e6fcd07792586309ddb4a6a1563b" Dec 04 10:44:57 crc kubenswrapper[4943]: I1204 10:44:57.565077 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:44:57 crc kubenswrapper[4943]: E1204 10:44:57.565649 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.158444 4943 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w"] Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.159893 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.161834 4943 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.171504 4943 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.174050 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w"] Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.312512 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/229ed355-6431-4a46-a20d-3d02888ca078-secret-volume\") pod \"collect-profiles-29414085-xpm8w\" (UID: \"229ed355-6431-4a46-a20d-3d02888ca078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.312617 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/229ed355-6431-4a46-a20d-3d02888ca078-config-volume\") pod \"collect-profiles-29414085-xpm8w\" (UID: \"229ed355-6431-4a46-a20d-3d02888ca078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.312788 4943 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z47nh\" (UniqueName: \"kubernetes.io/projected/229ed355-6431-4a46-a20d-3d02888ca078-kube-api-access-z47nh\") pod \"collect-profiles-29414085-xpm8w\" (UID: \"229ed355-6431-4a46-a20d-3d02888ca078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.414643 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z47nh\" (UniqueName: \"kubernetes.io/projected/229ed355-6431-4a46-a20d-3d02888ca078-kube-api-access-z47nh\") pod \"collect-profiles-29414085-xpm8w\" (UID: \"229ed355-6431-4a46-a20d-3d02888ca078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.414776 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/229ed355-6431-4a46-a20d-3d02888ca078-secret-volume\") pod \"collect-profiles-29414085-xpm8w\" (UID: \"229ed355-6431-4a46-a20d-3d02888ca078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.414832 4943 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/229ed355-6431-4a46-a20d-3d02888ca078-config-volume\") pod \"collect-profiles-29414085-xpm8w\" (UID: \"229ed355-6431-4a46-a20d-3d02888ca078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.416022 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/229ed355-6431-4a46-a20d-3d02888ca078-config-volume\") pod \"collect-profiles-29414085-xpm8w\" (UID: \"229ed355-6431-4a46-a20d-3d02888ca078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.421878 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/229ed355-6431-4a46-a20d-3d02888ca078-secret-volume\") pod \"collect-profiles-29414085-xpm8w\" (UID: \"229ed355-6431-4a46-a20d-3d02888ca078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.434527 4943 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z47nh\" (UniqueName: \"kubernetes.io/projected/229ed355-6431-4a46-a20d-3d02888ca078-kube-api-access-z47nh\") pod \"collect-profiles-29414085-xpm8w\" (UID: \"229ed355-6431-4a46-a20d-3d02888ca078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.485149 4943 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:00 crc kubenswrapper[4943]: I1204 10:45:00.896938 4943 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w"] Dec 04 10:45:01 crc kubenswrapper[4943]: I1204 10:45:01.853610 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" event={"ID":"229ed355-6431-4a46-a20d-3d02888ca078","Type":"ContainerStarted","Data":"3a1049e5b9f549dcaf82cafff93ee7e4f23a0cff8809671a15434fefa7a8868e"} Dec 04 10:45:02 crc kubenswrapper[4943]: I1204 10:45:02.860549 4943 generic.go:334] "Generic (PLEG): container finished" podID="229ed355-6431-4a46-a20d-3d02888ca078" containerID="27033fa75799da2769a97896dd3758a24e538c313a49596914cd6f450d193b43" exitCode=0 Dec 04 10:45:02 crc kubenswrapper[4943]: I1204 10:45:02.860871 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" event={"ID":"229ed355-6431-4a46-a20d-3d02888ca078","Type":"ContainerDied","Data":"27033fa75799da2769a97896dd3758a24e538c313a49596914cd6f450d193b43"} Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.192761 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.269218 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z47nh\" (UniqueName: \"kubernetes.io/projected/229ed355-6431-4a46-a20d-3d02888ca078-kube-api-access-z47nh\") pod \"229ed355-6431-4a46-a20d-3d02888ca078\" (UID: \"229ed355-6431-4a46-a20d-3d02888ca078\") " Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.269338 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/229ed355-6431-4a46-a20d-3d02888ca078-secret-volume\") pod \"229ed355-6431-4a46-a20d-3d02888ca078\" (UID: \"229ed355-6431-4a46-a20d-3d02888ca078\") " Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.269376 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/229ed355-6431-4a46-a20d-3d02888ca078-config-volume\") pod \"229ed355-6431-4a46-a20d-3d02888ca078\" (UID: \"229ed355-6431-4a46-a20d-3d02888ca078\") " Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.270234 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/229ed355-6431-4a46-a20d-3d02888ca078-config-volume" (OuterVolumeSpecName: "config-volume") pod "229ed355-6431-4a46-a20d-3d02888ca078" (UID: "229ed355-6431-4a46-a20d-3d02888ca078"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.274593 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/229ed355-6431-4a46-a20d-3d02888ca078-kube-api-access-z47nh" (OuterVolumeSpecName: "kube-api-access-z47nh") pod "229ed355-6431-4a46-a20d-3d02888ca078" (UID: "229ed355-6431-4a46-a20d-3d02888ca078"). InnerVolumeSpecName "kube-api-access-z47nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.278295 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/229ed355-6431-4a46-a20d-3d02888ca078-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "229ed355-6431-4a46-a20d-3d02888ca078" (UID: "229ed355-6431-4a46-a20d-3d02888ca078"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.371389 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z47nh\" (UniqueName: \"kubernetes.io/projected/229ed355-6431-4a46-a20d-3d02888ca078-kube-api-access-z47nh\") on node \"crc\" DevicePath \"\"" Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.371497 4943 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/229ed355-6431-4a46-a20d-3d02888ca078-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.371537 4943 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/229ed355-6431-4a46-a20d-3d02888ca078-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.872791 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" event={"ID":"229ed355-6431-4a46-a20d-3d02888ca078","Type":"ContainerDied","Data":"3a1049e5b9f549dcaf82cafff93ee7e4f23a0cff8809671a15434fefa7a8868e"} Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.872858 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414085-xpm8w" Dec 04 10:45:04 crc kubenswrapper[4943]: I1204 10:45:04.872869 4943 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a1049e5b9f549dcaf82cafff93ee7e4f23a0cff8809671a15434fefa7a8868e" Dec 04 10:45:09 crc kubenswrapper[4943]: I1204 10:45:09.565165 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:45:09 crc kubenswrapper[4943]: E1204 10:45:09.565885 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:45:10 crc kubenswrapper[4943]: I1204 10:45:10.610322 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/util/0.log" Dec 04 10:45:10 crc kubenswrapper[4943]: I1204 10:45:10.771176 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/util/0.log" Dec 04 10:45:10 crc kubenswrapper[4943]: I1204 10:45:10.771837 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/pull/0.log" Dec 04 10:45:10 crc kubenswrapper[4943]: I1204 10:45:10.772003 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/pull/0.log" Dec 04 10:45:10 crc kubenswrapper[4943]: I1204 10:45:10.928026 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/util/0.log" Dec 04 10:45:10 crc kubenswrapper[4943]: I1204 10:45:10.949018 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/extract/0.log" Dec 04 10:45:10 crc kubenswrapper[4943]: I1204 10:45:10.956067 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f839rbzw_7a3629de-83c4-4c6a-96e1-58d9bf06b57f/pull/0.log" Dec 04 10:45:11 crc kubenswrapper[4943]: I1204 10:45:11.107070 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/extract-utilities/0.log" Dec 04 10:45:11 crc kubenswrapper[4943]: I1204 10:45:11.253001 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/extract-content/0.log" Dec 04 10:45:11 crc kubenswrapper[4943]: I1204 10:45:11.259995 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/extract-utilities/0.log" Dec 04 10:45:11 crc kubenswrapper[4943]: I1204 10:45:11.308020 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/extract-content/0.log" Dec 04 10:45:11 crc kubenswrapper[4943]: I1204 10:45:11.422842 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/extract-content/0.log" Dec 04 10:45:11 crc kubenswrapper[4943]: I1204 10:45:11.423493 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/extract-utilities/0.log" Dec 04 10:45:11 crc kubenswrapper[4943]: I1204 10:45:11.595026 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/extract-utilities/0.log" Dec 04 10:45:11 crc kubenswrapper[4943]: I1204 10:45:11.774990 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-v55rk_f898614f-17cf-4133-9e79-08eff3ee3979/registry-server/0.log" Dec 04 10:45:11 crc kubenswrapper[4943]: I1204 10:45:11.836774 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/extract-utilities/0.log" Dec 04 10:45:11 crc kubenswrapper[4943]: I1204 10:45:11.856007 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/extract-content/0.log" Dec 04 10:45:11 crc kubenswrapper[4943]: I1204 10:45:11.881683 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/extract-content/0.log" Dec 04 10:45:12 crc kubenswrapper[4943]: I1204 10:45:12.026844 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/extract-content/0.log" Dec 04 10:45:12 crc kubenswrapper[4943]: I1204 10:45:12.036703 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/extract-utilities/0.log" Dec 04 10:45:12 crc kubenswrapper[4943]: I1204 10:45:12.324701 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/extract-utilities/0.log" Dec 04 10:45:12 crc kubenswrapper[4943]: I1204 10:45:12.329581 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-bwkhz_04c5f3f6-69da-4c89-801f-5e3d519c80ae/marketplace-operator/0.log" Dec 04 10:45:12 crc kubenswrapper[4943]: I1204 10:45:12.485374 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/extract-utilities/0.log" Dec 04 10:45:12 crc kubenswrapper[4943]: I1204 10:45:12.508493 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/extract-content/0.log" Dec 04 10:45:12 crc kubenswrapper[4943]: I1204 10:45:12.551386 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vbmdp_46236d49-e7f4-49aa-bd3e-e770c260553a/registry-server/0.log" Dec 04 10:45:12 crc kubenswrapper[4943]: I1204 10:45:12.588996 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/extract-content/0.log" Dec 04 10:45:12 crc kubenswrapper[4943]: I1204 10:45:12.711321 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/extract-content/0.log" Dec 04 10:45:12 crc kubenswrapper[4943]: I1204 10:45:12.714195 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/extract-utilities/0.log" Dec 04 10:45:12 crc kubenswrapper[4943]: I1204 10:45:12.825000 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c44sb_15b454e0-b590-41a9-b496-bf6577afbfed/registry-server/0.log" Dec 04 10:45:12 crc kubenswrapper[4943]: I1204 10:45:12.894534 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/extract-utilities/0.log" Dec 04 10:45:13 crc kubenswrapper[4943]: I1204 10:45:13.100878 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/extract-utilities/0.log" Dec 04 10:45:13 crc kubenswrapper[4943]: I1204 10:45:13.138088 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/extract-content/0.log" Dec 04 10:45:13 crc kubenswrapper[4943]: I1204 10:45:13.152419 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/extract-content/0.log" Dec 04 10:45:13 crc kubenswrapper[4943]: I1204 10:45:13.258873 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/extract-content/0.log" Dec 04 10:45:13 crc kubenswrapper[4943]: I1204 10:45:13.278498 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/extract-utilities/0.log" Dec 04 10:45:13 crc kubenswrapper[4943]: I1204 10:45:13.693887 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-blf86_facfc1e3-6604-4ec2-b320-b0508d57414b/registry-server/0.log" Dec 04 10:45:24 crc kubenswrapper[4943]: I1204 10:45:24.564828 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:45:24 crc kubenswrapper[4943]: E1204 10:45:24.565659 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:45:35 crc kubenswrapper[4943]: I1204 10:45:35.564970 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:45:35 crc kubenswrapper[4943]: E1204 10:45:35.565727 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:45:47 crc kubenswrapper[4943]: I1204 10:45:47.565082 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:45:47 crc kubenswrapper[4943]: E1204 10:45:47.566125 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:45:58 crc kubenswrapper[4943]: I1204 10:45:58.565068 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:45:58 crc kubenswrapper[4943]: E1204 10:45:58.565790 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:46:11 crc kubenswrapper[4943]: I1204 10:46:11.565818 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:46:11 crc kubenswrapper[4943]: E1204 10:46:11.567247 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:46:20 crc kubenswrapper[4943]: I1204 10:46:20.307588 4943 generic.go:334] "Generic (PLEG): container finished" podID="32ef8b0a-14c6-47b8-8052-d007e98c3e6c" containerID="b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b" exitCode=0 Dec 04 10:46:20 crc kubenswrapper[4943]: I1204 10:46:20.307707 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-ttvvb/must-gather-nvch6" event={"ID":"32ef8b0a-14c6-47b8-8052-d007e98c3e6c","Type":"ContainerDied","Data":"b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b"} Dec 04 10:46:20 crc kubenswrapper[4943]: I1204 10:46:20.308544 4943 scope.go:117] "RemoveContainer" containerID="b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b" Dec 04 10:46:20 crc kubenswrapper[4943]: I1204 10:46:20.647261 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ttvvb_must-gather-nvch6_32ef8b0a-14c6-47b8-8052-d007e98c3e6c/gather/0.log" Dec 04 10:46:26 crc kubenswrapper[4943]: I1204 10:46:26.564977 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:46:26 crc kubenswrapper[4943]: E1204 10:46:26.565963 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:46:30 crc kubenswrapper[4943]: I1204 10:46:30.084789 4943 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-ttvvb/must-gather-nvch6"] Dec 04 10:46:30 crc kubenswrapper[4943]: I1204 10:46:30.086146 4943 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-ttvvb/must-gather-nvch6" podUID="32ef8b0a-14c6-47b8-8052-d007e98c3e6c" containerName="copy" containerID="cri-o://56653d07a62fd3846df3301a93be11f647348b2e6623221a21431ae19f39a1ad" gracePeriod=2 Dec 04 10:46:30 crc kubenswrapper[4943]: I1204 10:46:30.092738 4943 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-ttvvb/must-gather-nvch6"] Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.238440 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ttvvb_must-gather-nvch6_32ef8b0a-14c6-47b8-8052-d007e98c3e6c/copy/0.log" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.239128 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttvvb/must-gather-nvch6" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.381554 4943 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-ttvvb_must-gather-nvch6_32ef8b0a-14c6-47b8-8052-d007e98c3e6c/copy/0.log" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.381854 4943 generic.go:334] "Generic (PLEG): container finished" podID="32ef8b0a-14c6-47b8-8052-d007e98c3e6c" containerID="56653d07a62fd3846df3301a93be11f647348b2e6623221a21431ae19f39a1ad" exitCode=143 Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.381920 4943 scope.go:117] "RemoveContainer" containerID="56653d07a62fd3846df3301a93be11f647348b2e6623221a21431ae19f39a1ad" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.381960 4943 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-ttvvb/must-gather-nvch6" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.397849 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcds8\" (UniqueName: \"kubernetes.io/projected/32ef8b0a-14c6-47b8-8052-d007e98c3e6c-kube-api-access-hcds8\") pod \"32ef8b0a-14c6-47b8-8052-d007e98c3e6c\" (UID: \"32ef8b0a-14c6-47b8-8052-d007e98c3e6c\") " Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.397915 4943 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/32ef8b0a-14c6-47b8-8052-d007e98c3e6c-must-gather-output\") pod \"32ef8b0a-14c6-47b8-8052-d007e98c3e6c\" (UID: \"32ef8b0a-14c6-47b8-8052-d007e98c3e6c\") " Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.402518 4943 scope.go:117] "RemoveContainer" containerID="b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.407610 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32ef8b0a-14c6-47b8-8052-d007e98c3e6c-kube-api-access-hcds8" (OuterVolumeSpecName: "kube-api-access-hcds8") pod "32ef8b0a-14c6-47b8-8052-d007e98c3e6c" (UID: "32ef8b0a-14c6-47b8-8052-d007e98c3e6c"). InnerVolumeSpecName "kube-api-access-hcds8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.457572 4943 scope.go:117] "RemoveContainer" containerID="56653d07a62fd3846df3301a93be11f647348b2e6623221a21431ae19f39a1ad" Dec 04 10:46:32 crc kubenswrapper[4943]: E1204 10:46:32.458058 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56653d07a62fd3846df3301a93be11f647348b2e6623221a21431ae19f39a1ad\": container with ID starting with 56653d07a62fd3846df3301a93be11f647348b2e6623221a21431ae19f39a1ad not found: ID does not exist" containerID="56653d07a62fd3846df3301a93be11f647348b2e6623221a21431ae19f39a1ad" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.458131 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56653d07a62fd3846df3301a93be11f647348b2e6623221a21431ae19f39a1ad"} err="failed to get container status \"56653d07a62fd3846df3301a93be11f647348b2e6623221a21431ae19f39a1ad\": rpc error: code = NotFound desc = could not find container \"56653d07a62fd3846df3301a93be11f647348b2e6623221a21431ae19f39a1ad\": container with ID starting with 56653d07a62fd3846df3301a93be11f647348b2e6623221a21431ae19f39a1ad not found: ID does not exist" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.458168 4943 scope.go:117] "RemoveContainer" containerID="b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b" Dec 04 10:46:32 crc kubenswrapper[4943]: E1204 10:46:32.458502 4943 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b\": container with ID starting with b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b not found: ID does not exist" containerID="b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.458541 4943 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b"} err="failed to get container status \"b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b\": rpc error: code = NotFound desc = could not find container \"b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b\": container with ID starting with b0454d009e623d75e20a8165b6a8651aded97a3e0d012db185db1bf3c97aee0b not found: ID does not exist" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.467579 4943 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32ef8b0a-14c6-47b8-8052-d007e98c3e6c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "32ef8b0a-14c6-47b8-8052-d007e98c3e6c" (UID: "32ef8b0a-14c6-47b8-8052-d007e98c3e6c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.498809 4943 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/32ef8b0a-14c6-47b8-8052-d007e98c3e6c-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.498857 4943 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcds8\" (UniqueName: \"kubernetes.io/projected/32ef8b0a-14c6-47b8-8052-d007e98c3e6c-kube-api-access-hcds8\") on node \"crc\" DevicePath \"\"" Dec 04 10:46:32 crc kubenswrapper[4943]: I1204 10:46:32.575472 4943 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32ef8b0a-14c6-47b8-8052-d007e98c3e6c" path="/var/lib/kubelet/pods/32ef8b0a-14c6-47b8-8052-d007e98c3e6c/volumes" Dec 04 10:46:37 crc kubenswrapper[4943]: I1204 10:46:37.565617 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:46:37 crc kubenswrapper[4943]: E1204 10:46:37.566042 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:46:52 crc kubenswrapper[4943]: I1204 10:46:52.570790 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:46:52 crc kubenswrapper[4943]: E1204 10:46:52.572939 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:47:07 crc kubenswrapper[4943]: I1204 10:47:07.565632 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:47:07 crc kubenswrapper[4943]: E1204 10:47:07.566837 4943 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-kswzd_openshift-machine-config-operator(7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501)\"" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" Dec 04 10:47:20 crc kubenswrapper[4943]: I1204 10:47:20.565539 4943 scope.go:117] "RemoveContainer" containerID="a711713f01b819135437895134887e0be672b01522668ee29b509ac434e35520" Dec 04 10:47:21 crc kubenswrapper[4943]: I1204 10:47:21.706671 4943 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" event={"ID":"7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501","Type":"ContainerStarted","Data":"e84c11251667ffc4155e21bbfe946eb7eeb4b38bf6c3a128c123da5efefb03ac"} Dec 04 10:49:48 crc kubenswrapper[4943]: I1204 10:49:48.629249 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:49:48 crc kubenswrapper[4943]: I1204 10:49:48.629712 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 10:50:18 crc kubenswrapper[4943]: I1204 10:50:18.629151 4943 patch_prober.go:28] interesting pod/machine-config-daemon-kswzd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 10:50:18 crc kubenswrapper[4943]: I1204 10:50:18.629746 4943 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-kswzd" podUID="7a5dd4d3-bf07-4a76-a0fd-c0cfd97a0501" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114263616024452 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114263617017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114257226016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114257226015462 5ustar corecore